Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118566
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorTan, H-
dc.creatorXu, C-
dc.creatorLi, J-
dc.creatorZhang, Y-
dc.creatorFang, Z-
dc.creatorChen, Z-
dc.creatorLai, B-
dc.date.accessioned2026-04-24T03:01:29Z-
dc.date.available2026-04-24T03:01:29Z-
dc.identifier.issn2162-237X-
dc.identifier.urihttp://hdl.handle.net/10397/118566-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights© 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication H. Tan et al., 'HICL: Hashtag-Driven In-Context Learning for Social Media Natural Language Understanding,' in IEEE Transactions on Neural Networks and Learning Systems, vol. 36, no. 4, pp. 7037-7050, April 2025 is available at https://doi.org/10.1109/TNNLS.2024.3384987.en_US
dc.subjectIn-context learning (ICL)en_US
dc.subjectNatural language processingen_US
dc.subjectPretrained language modelen_US
dc.subjectSocial mediaen_US
dc.titleHICL : hashtag-driven in-context learning for social media natural language understandingen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage7037-
dc.identifier.epage7050-
dc.identifier.volume36-
dc.identifier.issue4-
dc.identifier.doi10.1109/TNNLS.2024.3384987-
dcterms.abstractNatural language understanding (NLU) is integral to various social media applications. However, the existing NLU models rely heavily on context for semantic learning, resulting in compromised performance when faced with short and noisy social media content. To address this issue, we leverage in-context learning (ICL), wherein language models learn to make inferences by conditioning on a handful of demonstrations to enrich the context and propose a novel hashtag-driven ICL (HICL) framework. Concretely, we pretrain a model #Encoder, which employs #hashtags (user-annotated topic labels) to drive BERT-based pretraining through contrastive learning. Our objective here is to enable #Encoder to gain the ability to incorporate topic-related semantic information, which allows it to retrieve topic-related posts to enrich contexts and enhance social media NLU with noisy contexts. To further integrate the retrieved context with the source text, we employ a gradient-based method to identify trigger terms useful in fusing information from both sources. For empirical studies, we collected 45 M tweets to set up an in-context NLU benchmark, and the experimental results on seven downstream tasks show that HICL substantially advances the previous state-of-the-art results. Furthermore, we conducted an extensive analysis and found that the following hold: 1) combining source input with a top-retrieved post from #Encoder is more effective than using semantically similar posts and 2) trigger words can largely benefit in merging context from the source and retrieved posts.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE transactions on neural networks and learning systems, Apr. 2025, v. 36, no. 4, p. 7037-7050-
dcterms.isPartOfIEEE transactions on neural networks and learning systems-
dcterms.issued2025-04-
dc.identifier.scopus2-s2.0-105002373430-
dc.identifier.pmid38619957-
dc.identifier.eissn2162-2388-
dc.description.validate202604 bcjz-
dc.description.oaAccepted Manuscripten_US
dc.identifier.SubFormIDG001535/2025-12en_US
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported in part by the Research Grants Council of the Hong Kong Special Administrative Region, China, under Project PolyU/25200821; in part by the Innovation and Technology Fund under Project PRP/047/22FX; in part by the National Natural Science Foundation of China Young Scientists Fund under Grant 62006203; in part by the National Natural Science Foundation of China under Grant 62372220; and in part by the China Computer Federation-Baidu Open Research Fund under Grant 2021PP15002000.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
dc.relation.rdatahttps://github.com/albertan017/HICL-
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Tan_Hicl_Hashtag-driven_In-context.pdfPre-Published version2.72 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

2
Citations as of May 8, 2026

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.