Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116209
DC FieldValueLanguage
dc.contributorDepartment of Civil and Environmental Engineeringen_US
dc.creatorYu, Jen_US
dc.creatorWang, Yen_US
dc.creatorMa, Wen_US
dc.date.accessioned2025-12-02T03:29:26Z-
dc.date.available2025-12-02T03:29:26Z-
dc.identifier.issn1366-5545en_US
dc.identifier.urihttp://hdl.handle.net/10397/116209-
dc.language.isoenen_US
dc.publisherPergamon Pressen_US
dc.subjectBus bunchingen_US
dc.subjectControl strategyen_US
dc.subjectDeep reinforcement learningen_US
dc.subjectDynamic holdingen_US
dc.subjectLarge language modelen_US
dc.titleLarge language model-enhanced reinforcement learning for generic bus holding control strategiesen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume200en_US
dc.identifier.doi10.1016/j.tre.2025.104142en_US
dcterms.abstractBus holding control is a widely-adopted strategy for maintaining stability and improving the operational efficiency of bus systems. Traditional model-based methods often face challenges with the low accuracy of bus state prediction and passenger demand estimation. In contrast, Reinforcement Learning (RL), as a data-driven approach, has demonstrated great potential in formulating bus holding strategies. RL determines the optimal control strategies in order to maximize the cumulative reward, which reflects the overall control goals. However, translating sparse and delayed control goals in real-world tasks into dense and real-time rewards for RL is challenging, normally requiring extensive manual trial-and-error. In view of this, this study introduces an automatic reward generation paradigm by leveraging the in-context learning and reasoning capabilities of Large Language Models (LLMs). This new paradigm, termed the LLM-enhanced RL, comprises several LLM-based modules: reward initializer, reward modifier, performance analyzer, and reward refiner. These modules cooperate to initialize and iteratively improve the reward function according to the feedback from training and test results for the specified RL-based task. Ineffective reward functions generated by the LLM are filtered out to ensure the stable evolution of the RL agents’ performance over iterations. To evaluate the feasibility of the proposed LLM-enhanced RL paradigm, it is applied to extensive bus holding control scenarios that vary in the number of bus lines, stops, and passenger demand. The results demonstrate the superiority, generalization capability, and robustness of the proposed paradigm compared to vanilla RL strategies, the LLM-based controller, physics-based feedback controllers, and optimization-based controllers. This study sheds light on the great potential of utilizing LLMs in various smart mobility applications.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationTransportation research. Part E, Logistics and transportation review, Aug. 2025, v. 200, 104142en_US
dcterms.isPartOfTransportation research. Part E, Logistics and transportation reviewen_US
dcterms.issued2025-08-
dc.identifier.scopus2-s2.0-105005938308-
dc.identifier.eissn1878-5794en_US
dc.identifier.artn104142en_US
dc.description.validate202512 bcjzen_US
dc.description.oaNot applicableen_US
dc.identifier.SubFormIDG000407/2025-11-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThe work described in this paper is supported by the Innovation and Technology Fund - Mainland-Hong Kong Joint Funding Scheme (ITF-MHKJFS) (Project No. MHP/150/22), grants from the Research Grants Council of the Hong Kong Special Administrative Region, China (Project No. PolyU/15206322 and PolyU/15227424), and a grant from the Otto Poon Charitable Foundation Smart Cities Research Institute, The Hong Kong Polytechnic University (CD06). The contents of this article reflect the views of the authors, who are responsible for the facts and accuracy of the information presented herein.en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2028-08-31en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2028-08-31
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

4
Citations as of Apr 3, 2026

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.