Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/88789
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorSchool of Nursing-
dc.creatorJiang, DZ-
dc.creatorHuang, DM-
dc.creatorSong, YY-
dc.creatorWu, KC-
dc.creatorLu, HK-
dc.creatorLiu, QQ-
dc.creatorZhou, T-
dc.date.accessioned2020-12-22T01:07:58Z-
dc.date.available2020-12-22T01:07:58Z-
dc.identifier.urihttp://hdl.handle.net/10397/88789-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rightsThis work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/en_US
dc.rightsThe following publication Jiang, D. Z., Huang, D. M., Song, Y. Y., Wu, K. C., Lu, H. K., Liu, Q. Q., & Zhou, T. (2020). An audio data representation for traffic acoustic scene recognition. IEEE Access, 8, 177863-177873 is available at https://dx.doi.org/10.1109/ACCESS.2020.3027474en_US
dc.subjectAcousticsen_US
dc.subjectFeature extractionen_US
dc.subjectSpectrogramen_US
dc.subjectTransformsen_US
dc.subjectHistogramsen_US
dc.subjectTime-Frequency analysisen_US
dc.subjectVisualizationen_US
dc.subjectFeature extractionen_US
dc.subjectAcoustic scene recognitionen_US
dc.subjectTransportationen_US
dc.subjectAcoustic materialen_US
dc.titleAn audio data representation for traffic acoustic scene recognitionen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage177863-
dc.identifier.epage177873-
dc.identifier.volume8-
dc.identifier.doi10.1109/ACCESS.2020.3027474-
dcterms.abstractAcoustic scene recognition (ASR), recognizing acoustic environments given an audio recording of the scene, has a wide range of applications, e.g. robotic navigation and audio forensic. However, ASR remains challenging mainly due to the difficulty of representing audio data. In this article, we focus on traffic acoustic data. Traffic acoustic sense recognition provides complementary information to visual information of the scene; for example, it can be used to verify the visual perception result. The acoustic analysis and recognition, in consideration of its simple and convenient, can effectively enhance the perception ability which only applies visual information. We propose an audio data representation method to improve the traffic acoustic scene recognition accuracy. The proposed method employs the constant Q transform (CQT) and histogram of gradient (HOG) to transfer the one-dimensional audio signals into a time-frequency representation. We also propose two data representation mechanisms, called global and local feature selections, in order to select features that are able to describe the shape of time-frequency structures. We finally exploit the least absolute shrinkage and selection operator (LASSO) technique to further improve the recognition accuracy, by further selecting the most representative information for the recognition. We implemented extensive experiments, and the results show that the proposed method is effective, significantly outperforming the state-of-the-art methods.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE access, . . 2020, , v. 8, p. 177863-177873-
dcterms.isPartOfIEEE access-
dcterms.issued2020-
dc.identifier.isiWOS:000576244600001-
dc.identifier.eissn2169-3536-
dc.description.validate202012 bcrc-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.pubStatusPublisheden_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Jiang_Audio_Data_Representation.pdf1.64 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

42
Last Week
0
Last month
Citations as of May 19, 2024

Downloads

23
Citations as of May 19, 2024

SCOPUSTM   
Citations

11
Citations as of May 17, 2024

WEB OF SCIENCETM
Citations

8
Citations as of May 16, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.