Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/117186
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Data Science and Artificial Intelligenceen_US
dc.contributorDepartment of Computingen_US
dc.creatorSong, Zen_US
dc.creatorZhang, Sen_US
dc.creatorChou, Yen_US
dc.creatorWu, Jen_US
dc.creatorLi, Hen_US
dc.date.accessioned2026-02-06T02:07:13Z-
dc.date.available2026-02-06T02:07:13Z-
dc.identifier.issn2162-237Xen_US
dc.identifier.urihttp://hdl.handle.net/10397/117186-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication Z. Song, S. Zhang, Y. Chou, J. Wu and H. Li, "IML-Spikeformer: Input-Aware Multilevel Spiking Transformer for Speech Processing," in IEEE Transactions on Neural Networks and Learning Systems, vol. 37, no. 3, pp. 1377-1389, March 2026 is available at https://doi.org/10.1109/TNNLS.2025.3615971.en_US
dc.subjectNeuromorphic auditory processingen_US
dc.subjectSpeech recognitionen_US
dc.subjectSpiking neural networks (SNNs)en_US
dc.subjectSpiking trans- formeren_US
dc.titleIML-Spikeformer : input-aware multilevel spiking transformer for speech processingen_US
dc.typeJournal/Magazine Articleen_US
dc.description.otherinformationTitle on author's file: IML-Spikeformer: Input-aware Multi-Level Spiking Transformer for Speech Processingen_US
dc.identifier.spage1377en_US
dc.identifier.epage1389en_US
dc.identifier.volume37en_US
dc.identifier.issue3en_US
dc.identifier.doi10.1109/TNNLS.2025.3615971en_US
dcterms.abstractSpiking neural networks (SNNs), inspired by biological neural mechanisms, represent a promising neuromorphic computing paradigm that offers energy-efficient alternatives to traditional artificial neural networks (ANNs). Despite proven effectiveness, SNN architectures have struggled to achieve competitive performance on large-scale speech processing tasks. Two key challenges hinder progress: 1) the high computational overhead during training caused by multitimestep spike firing and 2) the absence of large-scale SNN architectures tailored to speech processing tasks. To overcome the issues, we introduce the input-aware multilevel spikeformer (IML-Spikeformer), a spiking transformer architecture specifically designed for large-scale speech processing. Central to our design is the input-aware multilevel spike (IMLS) mechanism, which simulates multitimestep spike firing within a single timestep using an adaptive, input-aware thresholding scheme. IML-Spikeformer further integrates a reparameterized spiking self-attention (RepSSA) module with a hierarchical decay mask (HDM), forming the HD-RepSSA module. This module enhances the precision of attention maps and enables modeling of multiscale temporal dependencies in speech signals. Experiments demonstrate that IML-Spikeformer achieves word error rates (WERs) of 6.0% on AiShell-1 and 3.4% on Librispeech-960, comparable to conventional ANN transformers while reducing theoretical inference energy consumption by 4.64X and 4.32X, , respectively. IML-Spikeformer marks an advance of scalable SNN architectures for large-scale speech processing in both task performance and energy efficiency. Our source code and model checkpoints are publicly available at github.com/Pooookeman/IML-Spikeformeren_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on neural networks and learning systems, Mar. 2026, v. 37, no. 3, p. 1377-1389en_US
dcterms.isPartOfIEEE transactions on neural networks and learning systemsen_US
dcterms.issued2026-03-
dc.identifier.eissn2162-2388en_US
dc.description.validate202602 bcchen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera4305-
dc.identifier.SubFormID52560-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported in part by the National Natural Science Foundation of China under Grant 62271432 and Grant 62306259; in part by Shenzhen Science and Technology Program (Shenzhen Key Laboratory) under Grant ZDSYS20230626091302006; in part by the Program for Guangdong Introducing Innovative and Entrepreneurial Teams, under Grant 2023ZT10X044; and in part by the Research Grants Council of the 5 Hong Kong, SAR under Grant C5052-23G, Grant PolyU15217424, and Grant PolyU25216423.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Song_IML_Spikeformer_Input.pdfPre-Published version882.91 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.