Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/101455
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorXu, Cen_US
dc.creatorLi, Jen_US
dc.date.accessioned2023-09-18T02:26:39Z-
dc.date.available2023-09-18T02:26:39Z-
dc.identifier.urihttp://hdl.handle.net/10397/101455-
dc.descriptionThe 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, December 7–11, 2022en_US
dc.language.isoenen_US
dc.publisherAssociation for Computational Linguisticsen_US
dc.rights©2022 Association for Computational Linguistics.en_US
dc.rightsMaterials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License. (https://creativecommons.org/licenses/by/4.0/)en_US
dc.rightsThe following publication Chunpu Xu and Jing Li. 2022. Borrowing Human Senses: Comment-Aware Self-Training for Social Media Multimodal Classification. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 5644–5656, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics is available at https://doi.org/10.18653/v1/2022.emnlp-main.381.en_US
dc.titleBorrowing human senses : comment-aware self-training for social media multimodal classificationen_US
dc.typeConference Paperen_US
dc.identifier.spage5644en_US
dc.identifier.epage5656en_US
dc.identifier.doi10.18653/v1/2022.emnlp-main.381en_US
dcterms.abstractSocial media is daily creating massive multimedia content with paired image and text, presenting the pressing need to automate the vision and language understanding for various multimodal classification tasks. Compared to the commonly researched visual-lingual data, social media posts tend to exhibit more implicit image-text relations. To better glue the cross-modal semantics therein, we capture hinting features from user comments, which are retrieved via jointly leveraging visual and lingual similarity. Afterwards, the classification tasks are explored via self-training in a teacher-student framework, motivated by the usually limited labeled data scales in existing benchmarks. Substantial experiments are conducted on four multimodal social media benchmarks for image-text relation classification, sarcasm detection, sentiment classification, and hate speech detection. The results show that our method further advances the performance of previous state-of-the-art models, which do not employ comment modeling or self-training.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7-11 Dec. 2022, p. 5466 - 5656en_US
dcterms.issued2022-
dc.identifier.ros2022002839-
dc.relation.conferenceConference on Empirical Methods in Natural Language Processing [EMNLP]-
dc.description.validate202309 bcww-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberCDCF_2022-2023-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Xu_Borrowing_Human_Senses.pdf1.97 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

142
Last Week
12
Last month
Citations as of Nov 9, 2025

Downloads

67
Citations as of Nov 9, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.