Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/102663
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.contributorSchool of Optometryen_US
dc.contributorDepartment of Rehabilitation Sciencesen_US
dc.creatorWang, Ken_US
dc.creatorCheung, MKMen_US
dc.creatorZhang, Yen_US
dc.creatorYang, Cen_US
dc.creatorChen, PQen_US
dc.creatorFu, EYen_US
dc.creatorNgai, Gen_US
dc.date.accessioned2023-11-06T01:14:30Z-
dc.date.available2023-11-06T01:14:30Z-
dc.identifier.isbn979-8-4007-0108-5en_US
dc.identifier.urihttp://hdl.handle.net/10397/102663-
dc.descriptionThe 31st ACM International Conference on Multimedia, Ottawa ON Canada, 29 October 2023 - 3 November 2023en_US
dc.language.isoenen_US
dc.publisherAssociation for Computing Machineryen_US
dc.rights© 2023 Copyright held by the owner/author(s). Publication rights licensed to ACM. This is the author's version of the work. It is posted here for your personal use. Not for redistribution. The definitive Version of Record was published in MM '23: Proceedings of the 31st ACM International Conference on Multimedia, https://doi.org/10.1145/3581783.3612870.en_US
dc.subjectBackchannel detectionen_US
dc.subjectAttention modelsen_US
dc.subjectVisual cuesen_US
dc.titleUnveiling subtle cues : backchannel detection using temporal multimodal attention networksen_US
dc.typeConference Paperen_US
dc.identifier.spage9586en_US
dc.identifier.epage9590en_US
dc.identifier.doi10.1145/3581783.3612870en_US
dcterms.abstractAutomatic detection of backchannel has great potential to enhance artificial mediators, which indicate listeners' attention and agreement in human communication. It is often expressed by subtle non-verbal cues that occur briefly and sparsely. Focusing on identifying and locating these subtle cues (i.e., their occurrence moment and the involved body parts), this paper proposes a novel approach for backchannel detection. In particular, our model utilizes temporal- and modality-attention modules to determine and lead the model to pay more attention to both the indicative moment and the accompanying body parts at that specific time. It achieves an accuracy of 68.6% on the testing set in MultiMediate'23 backchannel detection challenge, outperforming the counterparts. Furthermore, we conducted an ablation study to thoroughly understand the contributions of our model. This study underscores the effectiveness of our selection of modality inputs and the importance of the two attention modules in our model.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn MM '23: Proceedings of the 31st ACM International Conference on Multimedia, p. 9586-9590. New York, NY: Association for Computing Machinery, 2023en_US
dcterms.issued2023-
dc.relation.ispartofbookMM '23 : Proceedings of the 31st ACM International Conference on Multimediaen_US
dc.relation.conferenceACM International Conference on Multimedia [MM]en_US
dc.description.validate202311 bcchen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera2504-
dc.identifier.SubFormID47794-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Wang_Unveiling_Subtle_Cues.pdfPre-Published version6.47 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

102
Citations as of Apr 14, 2025

Downloads

145
Citations as of Apr 14, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.