Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/115490
DC FieldValueLanguage
dc.contributorDepartment of Logistics and Maritime Studies-
dc.creatorCao, Zen_US
dc.creatorSun, Qen_US
dc.creatorWang, Wen_US
dc.creatorWang, Sen_US
dc.date.accessioned2025-10-02T02:22:27Z-
dc.date.available2025-10-02T02:22:27Z-
dc.identifier.issn0968-090Xen_US
dc.identifier.urihttp://hdl.handle.net/10397/115490-
dc.language.isoenen_US
dc.publisherElsevier Ltden_US
dc.subjectBerth allocationen_US
dc.subjectDry bulk export terminalen_US
dc.subjectMulti-agent reinforcement learningen_US
dc.subjectNavigation channelen_US
dc.subjectPort operationen_US
dc.subjectShip schedulingen_US
dc.titleBerth allocation in dry bulk export terminals with channel restrictionsen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume179en_US
dc.identifier.doi10.1016/j.trc.2025.105263en_US
dcterms.abstractEfficient berth allocation (BA) is critical to port management, as berthing time and location directly impact operational efficiency. In dry bulk export terminals, the BA problem becomes more complex due to deballasting delays and pre-deballasting procedures, particularly under restrictive channel conditions. Terminal operators must balance pre-deballasting requirements with timely berthing to minimize delays. To address these challenges, we formulate the BA problem as a dynamic program, enabling sequential decision-making for each ship at every stage. To address the extensive state-action space, we propose a hierarchical decision framework that divides each stage into four planning-level substages and one scheduling-level substage, each handled by a dedicated agent. The planning level determines berthing positions and ship sequence, while the scheduling level coordinates berthing, channel access, and deballasting timelines based on planning outcomes. We propose a Planning by Reinforcement Learning and Scheduling by Optimization (PRLSO) approach, where agents employ either reinforcement learning (RL) or optimization, depending on substage characteristics. By confining RL-based agents to a reduced decision space, we significantly reduce training complexity. Following this, the remaining scheduling problem is solved on a reduced scale free from computational challenge. Experimental results show that the proposed method generates high-quality solutions in near real-time, even for large-scale instances. The framework also improves training efficiency and supports industrial-scale implementation.-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationTransportation research. Part C, Emerging technologies, Oct. 2025, v. 179, 105263en_US
dcterms.isPartOfTransportation research. Part C, Emerging technologiesen_US
dcterms.issued2025-10-
dc.identifier.scopus2-s2.0-105010702405-
dc.identifier.eissn1879-2359en_US
dc.identifier.artn105263en_US
dc.description.validate202510 bcch-
dc.description.oaNot applicableen_US
dc.identifier.SubFormIDG000156/2025-08-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported by the National Natural Science Foundation of China [Grant No. 72371221, 72371143]. The authors would like to express gratitude to the PolyU Maritime Data and Sustainable Development Centre (PMDC) for its invaluable support in providing data and equipment.en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2027-10-31en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2027-10-31
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.