Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/107478
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorSong, Zen_US
dc.creatorWu, Jen_US
dc.creatorZhang, Men_US
dc.creatorShou, MZen_US
dc.creatorLi, Hen_US
dc.date.accessioned2024-06-27T01:33:35Z-
dc.date.available2024-06-27T01:33:35Z-
dc.identifier.isbn979-8-3503-4485-1en_US
dc.identifier.urihttp://hdl.handle.net/10397/107478-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication Z. Song, J. Wu, M. Zhang, M. Z. Shou and H. Li, "Spiking-Leaf: A Learnable Auditory Front-End for Spiking Neural Networks," ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Korea, Republic of, 2024, pp. 226-230 is available at https://doi.org/10.1109/ICASSP48485.2024.10446789.en_US
dc.subjectLearnable audio front-enden_US
dc.subjectSpeech recognitionen_US
dc.subjectSpike encodingen_US
dc.subjectSpiking neural networksen_US
dc.titleSpiking-leaf : a learnable auditory front-end for spiking neural networksen_US
dc.typeConference Paperen_US
dc.identifier.spage226en_US
dc.identifier.epage230en_US
dc.identifier.doi10.1109/ICASSP48485.2024.10446789en_US
dcterms.abstractBrain-inspired spiking neural networks (SNNs) have demonstrated great potential for temporal signal processing. However, their performance in speech processing remains limited due to the lack of an effective auditory front-end. To address this limitation, we introduce Spiking-LEAF, a learnable auditory front-end meticulously designed for SNN-based speech processing. Spiking-LEAF combines a learnable filter bank with a novel two-compartment spiking neuron model called IHC-LIF. The IHC-LIF neurons draw inspiration from the structure of inner hair cells (IHC) and they leverage segregated dendritic and somatic compartments to effectively capture multi-scale temporal dynamics of speech signals. Additionally, the IHC-LIF neurons incorporate the lateral feedback mechanism along with spike regularization loss to enhance spike encoding efficiency. On keyword spotting and speaker identification tasks, the proposed Spiking-LEAF outperforms both SOTA spiking auditory front-ends and conventional real-valued acoustic features in terms of classification accuracy, noise robustness, and encoding efficiency.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitation2024 IEEE International Conference on Acoustics, Speech,and Signal Processing : Proceedings : 14-19 April 2024, COEX, Seoul, Korea, p. 226-230en_US
dcterms.issued2024-
dc.relation.conferenceInternational Conference on Acoustics, Speech, and Signal Processing [ICASSP]-
dc.description.validate202406 bcch-
dc.description.oaAuthor’s Originalen_US
dc.identifier.FolderNumbera2887-
dc.identifier.SubFormID48649-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextNational Natural Science Foundation of Chinaen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AO)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Song_Spiking-leaf_Learnable_Auditory.pdfPreprint version1.19 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Author’s Original
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

4
Citations as of Jun 30, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.