Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/107875
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorLi, Xen_US
dc.creatorLi, Jen_US
dc.date.accessioned2024-07-15T07:55:28Z-
dc.date.available2024-07-15T07:55:28Z-
dc.identifier.isbn979-8-89176-114-8en_US
dc.identifier.urihttp://hdl.handle.net/10397/107875-
dc.descriptionThe 2024 Conference of the North American Chapter of the Association for Computational Linguistics, June 16-21, 2024, Mexico Cityen_US
dc.language.isoenen_US
dc.publisherAssociation for Computational Linguistics (ACL)en_US
dc.rights© 2024 Association for Computational Linguisticsen_US
dc.rightsMaterials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Xianming Li and Jing Li. 2024. BeLLM: Backward Dependency Enhanced Large Language Model for Sentence Embeddings. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 792–804, Mexico City, Mexico. Association for Computational Linguistics is available at https://aclanthology.org/2024.naacl-long.45/.en_US
dc.titleBeLLM : backward dependency enhanced large language model for sentence embeddingsen_US
dc.typeConference Paperen_US
dc.identifier.spage792en_US
dc.identifier.epage804en_US
dc.identifier.volume1en_US
dcterms.abstractSentence embeddings are crucial in measuring semantic similarity. Most recent studies employed large language models (LLMs) to learn sentence embeddings. Existing LLMs mainly adopted autoregressive architecture without explicit backward dependency modeling. Therefore, we examined the effects of backward dependencies in LLMs for semantic similarity measurements. Concretely, we propose a novel model: backward dependency enhanced large language model (BeLLM). It learns sentence embeddings via transforming specific attention layers from uni- to bi-directional. We extensively experiment across various semantic textual similarity (STS) tasks and downstream applications. BeLLM achieves state-of-the-art performance in varying scenarios. It shows that autoregressive LLMs benefit from backward dependencies for sentence embeddings.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), p. 792–804, Mexico City, Mexico. Association for Computational Linguisticsen_US
dcterms.issued2024-
dc.relation.conferenceConference of the North American Chapter of the Association for Computational Linguistics [NAACL]en_US
dc.description.validate202407 bcwhen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera3031-
dc.identifier.SubFormID49239-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
2024.naacl-long.45.pdf789.83 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

6
Citations as of Jul 21, 2024

Downloads

1
Citations as of Jul 21, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.