Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116451
DC FieldValueLanguage
dc.contributorDepartment of Aeronautical and Aviation Engineeringen_US
dc.creatorLiu, Qen_US
dc.creatorLi, Fen_US
dc.creatorNg, KKHen_US
dc.creatorHan, Jen_US
dc.creatorFeng, Sen_US
dc.date.accessioned2025-12-30T05:47:41Z-
dc.date.available2025-12-30T05:47:41Z-
dc.identifier.issn0957-4174en_US
dc.identifier.urihttp://hdl.handle.net/10397/116451-
dc.language.isoenen_US
dc.publisherPergamon Pressen_US
dc.subjectAccident investigationen_US
dc.subjectChain of thoughten_US
dc.subjectGeneral aviationen_US
dc.subjectHFACSen_US
dc.subjectLarge language modelsen_US
dc.subjectWitness narrativesen_US
dc.titleAccident investigation via LLMs reasoning : HFACS-guided chain-of-thoughts enhance general aviation safetyen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume269en_US
dc.identifier.doi10.1016/j.eswa.2025.126422en_US
dcterms.abstractAviation accident investigation is crucial for preventing future accidents. However, traditional investigations in general aviation (GA) are expert-dependent and time-consuming. This study explores the potential of large language models (LLMs) to expedite this process by inferring human errors from witness narratives. Despite their promise, LLMs still struggle with domain-specific reasoning. To address this, we proposed a novel HFACS-CoT prompt that integrates the Human Factors Analysis and Classification System (HFACS) with Chain of Thought (CoT) reasoning, guiding LLMs to infer the pilot's unsafe acts and preconditions in a multi-step, two-stage process. HFACS-CoT+ further refines this prompt by sequentially guiding LLMs through each step and replacing textual instructions with programmatic logic statements. A new HFACS-labeled GA accident dataset was developed to support GA safety research as well as validate our proposed prompts. Using GPT-4o with the selected dataset, we found that HFACS-CoT significantly enhances LLMs’ ability to infer human errors, outperforming basic zero-shot, basic few-shot, auto-CoT and plan-and-solve prompts. HFACS-CoT+ further improves inference of preconditions and addresses deficiencies in answering logic. Moreover, comparative evaluations indicate that LLM surpass human experts in inferring certain human errors. This study highlights the benefits of integrating domain knowledge into prompt design and the potential of LLMs in GA accident investigations.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationExpert systems with applications, 15 Apr. 2025, v. 269, 126422en_US
dcterms.isPartOfExpert systems with applicationsen_US
dcterms.issued2025-04-15-
dc.identifier.scopus2-s2.0-85214674440-
dc.identifier.eissn1873-6793en_US
dc.identifier.artn126422en_US
dc.description.validate202512 bchyen_US
dc.description.oaNot applicableen_US
dc.identifier.SubFormIDG000558/2025-12-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThe study is partially supported by The Hong Kong Polytechnic University ( P0038933 and P0038827 ), by Research Centre Data Science AI ( P0042711 ). In addition, great appreciation is especially given to the anonymous reviewers, whose constructive feedback and suggestions greatly contributed to the improvement of the manuscript.en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2027-04-15en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2027-04-15
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.