Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116648
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorHong, Men_US
dc.creatorZhang, CJen_US
dc.creatorYang, Len_US
dc.creatorSong, Yen_US
dc.creatorJiang, Den_US
dc.date.accessioned2026-01-09T03:07:29Z-
dc.date.available2026-01-09T03:07:29Z-
dc.identifier.urihttp://hdl.handle.net/10397/116648-
dc.descriptionThe 16th Asian Conference on Machine Learning, December 5-8, 2024, in Hanoi, Vietnamen_US
dc.language.isoenen_US
dc.publisherPMLR web siteen_US
dc.rights© 2024 M. Hong, C.J. Zhang, L. Yang, Y. Song & D. Jiang.en_US
dc.rightsPosted with permission of the author.en_US
dc.rightsThe following publication Hong, M., Zhang, C.J., Yang, L., SONG, Y. & Jiang, D.. (2025). InfantCryNet: A Data-driven Framework for Intelligent Analysis of Infant Cries. Proceedings of the 16th Asian Conference on Machine Learning, in Proceedings of Machine Learning Research 260:845-857 is available at https://proceedings.mlr.press/v260/hong25a.html.en_US
dc.subjectConvolutional neural networksen_US
dc.subjectInfant cry classificationen_US
dc.subjectModel compressionen_US
dc.titleInfantCryNet : a data-driven framework for intelligent analysis of infant criesen_US
dc.typeConference Paperen_US
dc.identifier.spage845en_US
dc.identifier.epage857en_US
dc.identifier.volume260en_US
dcterms.abstractUnderstanding the meaning of infant cries is a significant challenge for young parents in caring for their newborns. The presence of background noise and the lack of labeled data present practical challenges in developing systems that can detect crying and analyze its underlying reasons. In this paper, we present a novel data-driven framework, “InfantCryNet,” for accomplishing these tasks. To address the issue of data scarcity, we employ pre-trained audio models to incorporate prior knowledge into our model. We propose the use of statistical pooling and multi-head attention pooling techniques to extract features more effectively. Additionally, knowledge distillation and model quantization are applied to enhance model efficiency and reduce the model size, better supporting industrial deployment in mobile devices. Experiments on real-life datasets demonstrate the superior performance of the proposed framework, outperforming state-of-the-art baselines by 4.4% in classification accuracy. The model compression effectively reduces the model size by 7% without compromising performance and by up to 28% with only an 8% decrease in accuracy, offering practical insights for model selection and system design.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of the 16th Asian Conference on Machine Learning, 2024, v. 260, p. 845-857en_US
dcterms.issued2024-
dc.relation.conferenceAsian Conference on Machine Learning [ACML]en_US
dc.description.validate202601 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera4255a-
dc.identifier.SubFormID52470-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCopyright retained by authoren_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
hong25a.pdf2.99 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Show simple item record

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.