Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/106691
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Chinese and Bilingual Studiesen_US
dc.creatorKauf, Cen_US
dc.creatorIvanova, AAen_US
dc.creatorRambelli, Gen_US
dc.creatorChersoni, Een_US
dc.creatorShe, JSen_US
dc.creatorChowdhury, Zen_US
dc.creatorFedorenko, Een_US
dc.creatorLenci, Aen_US
dc.date.accessioned2024-06-03T02:11:32Z-
dc.date.available2024-06-03T02:11:32Z-
dc.identifier.urihttp://hdl.handle.net/10397/106691-
dc.language.isoenen_US
dc.publisherWiley-Blackwell Publishing, Inc.en_US
dc.rights© 2023 The Authors. Cognitive Science published by Wiley Periodicals LLC on behalf of Cognitive Science Society (CSS).en_US
dc.rightsThis is an open access article under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits use, distribution and reproduction in any medium, provided the original work is properly cited.en_US
dc.rightsThe following publication Kauf, C., Ivanova, A.A., Rambelli, G., Chersoni, E., She, J.S., Chowdhury, Z., Fedorenko, E. and Lenci, A. (2023), Event Knowledge in Large Language Models: The Gap Between the Impossible and the Unlikely. Cognitive Science, 47: e13386 is available at https://doi.org/10.1111/cogs.13386.en_US
dc.subjectArtificial neural networksen_US
dc.subjectGeneralized event knowledgeen_US
dc.subjectLanguage modelsen_US
dc.subjectPlausibilityen_US
dc.subjectSemanticsen_US
dc.subjectSyntaxen_US
dc.subjectTypicalityen_US
dc.subjectWorld knowledgeen_US
dc.titleEvent knowledge in large language models : the gap between the impossible and the unlikelyen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume47en_US
dc.identifier.issue11en_US
dc.identifier.doi10.1111/cogs.13386en_US
dcterms.abstractWord co-occurrence patterns in language corpora contain a surprising amount of conceptual knowledge. Large language models (LLMs), trained to predict words in context, leverage these patterns to achieve impressive performance on diverse semantic tasks requiring world knowledge. An important but understudied question about LLMs’ semantic abilities is whether they acquire generalized knowledge of common events. Here, we test whether five pretrained LLMs (from 2018's BERT to 2023's MPT) assign a higher likelihood to plausible descriptions of agent−patient interactions than to minimally different implausible versions of the same event. Using three curated sets of minimal sentence pairs (total n = 1215), we found that pretrained LLMs possess substantial event knowledge, outperforming other distributional language models. In particular, they almost always assign a higher likelihood to possible versus impossible events (The teacher bought the laptop vs. The laptop bought the teacher). However, LLMs show less consistent preferences for likely versus unlikely events (The nanny tutored the boy vs. The boy tutored the nanny). In follow-up analyses, we show that (i) LLM scores are driven by both plausibility and surface-level sentence features, (ii) LLM scores generalize well across syntactic variants (active vs. passive constructions) but less well across semantic variants (synonymous sentences), (iii) some LLM errors mirror human judgment ambiguity, and (iv) sentence plausibility serves as an organizing dimension in internal LLM representations. Overall, our results show that important aspects of event knowledge naturally emerge from distributional linguistic patterns, but also highlight a gap between representations of possible/impossible and likely/unlikely events.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationCognitive science, Nov. 2023, v. 47, e13386en_US
dcterms.isPartOfCognitive scienceen_US
dcterms.issued2023-11-
dc.identifier.scopus2-s2.0-85177808989-
dc.identifier.eissn1551-6709en_US
dc.identifier.artne13386en_US
dc.description.validate202405 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera2727a-
dc.identifier.SubFormID48136-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
dc.relation.rdatahttps://github.com/carina-kauf/lm-event-knowledgeen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Kauf_Event_Knowledge_Large.pdf1.77 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

5
Citations as of Jun 30, 2024

Downloads

4
Citations as of Jun 30, 2024

SCOPUSTM   
Citations

6
Citations as of Jun 21, 2024

WEB OF SCIENCETM
Citations

5
Citations as of Jun 27, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.