Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/113676
DC FieldValueLanguage
dc.contributorDepartment of Data Science and Artificial Intelligence-
dc.contributorDepartment of Computing-
dc.creatorHao, X-
dc.creatorMa, C-
dc.creatorYang, Q-
dc.creatorWu, J-
dc.creatorTan, KC-
dc.date.accessioned2025-06-17T07:40:49Z-
dc.date.available2025-06-17T07:40:49Z-
dc.identifier.issn2162-237X-
dc.identifier.urihttp://hdl.handle.net/10397/113676-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.subjectNeuromorphic computingen_US
dc.subjectNeuromorphic speech processingen_US
dc.subjectSpeech enhancement (SE)en_US
dc.subjectSpiking neural network (SNN)en_US
dc.titleToward ultralow-power neuromorphic speech enhancement with Spiking-FullSubNeten_US
dc.typeJournal/Magazine Articleen_US
dc.description.otherinformationTitle on author's file: Towards Ultra-Low-Power Neuromorphic Speech Enhancement with Spiking-FullSubNeten_US
dc.identifier.doi10.1109/TNNLS.2025.3566021-
dcterms.abstractSpeech enhancement (SE) is critical for improving speech intelligibility and quality in various audio devices. In recent years, deep learning-based methods have significantly improved SE performance, but they often come with a high computational cost, which is prohibitive for a large number of edge devices, such as headsets and hearing aids. This work proposes an ultralow-power SE system based on the brain-inspired spiking neural network (SNN) called Spiking-FullSubNet. Spiking-FullSubNet follows a full-band and subband fusioned approach to effectively capture both global and local spectral information. To enhance the efficiency of computationally expensive subband modeling, we introduce a frequency partitioning method inspired by the sensitivity profile of the human peripheral auditory system. Furthermore, we introduce a novel spiking neuron model that can dynamically control the input information integration and forgetting, enhancing the multiscale temporal processing capability of SNN, which is critical for speech denoising. Experiments conducted on the recent Intel Neuromorphic Deep Noise Suppression (N-DNS) Challenge dataset show that the Spiking-FullSubNet surpasses state-of-the-art (SOTA) methods by large margins in terms of both speech quality and energy efficiency metrics. Notably, our system won the championship of the Intel N-DNS Challenge (algorithmic track), opening up a myriad of opportunities for ultralow-power SE at the edge. Our source code and model checkpoints are publicly available at github.com/haoxiangsnr/spiking-fullsubnet-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationIEEE transactions on neural networks and learning systems, Date of Publication: 15 May 2025, Early Access, https://doi.org/10.1109/TNNLS.2025.3566021-
dcterms.isPartOfIEEE transactions on neural networks and learning systems-
dcterms.issued2025-
dc.identifier.scopus2-s2.0-105005554696-
dc.identifier.eissn2162-2388-
dc.description.validate202506 bcch-
dc.identifier.FolderNumbera3717ben_US
dc.identifier.SubFormID50839en_US
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextNational Natural Science Foundation of Chinaen_US
dc.description.pubStatusEarly releaseen_US
dc.date.embargo0000-00-00 (to be updated)en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 0000-00-00 (to be updated)
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.