Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/115393
DC FieldValueLanguage
dc.contributorDepartment of Industrial and Systems Engineering-
dc.contributorResearch Institute for Advanced Manufacturing-
dc.creatorSun, M-
dc.creatorDing, J-
dc.creatorZhao, Z-
dc.creatorChen, J-
dc.creatorHuang, GQ-
dc.creatorWang, L-
dc.date.accessioned2025-09-23T03:16:44Z-
dc.date.available2025-09-23T03:16:44Z-
dc.identifier.issn0736-5845-
dc.identifier.urihttp://hdl.handle.net/10397/115393-
dc.language.isoenen_US
dc.publisherPergamon Pressen_US
dc.subjectOut-of-orderen_US
dc.subjectDynamic schedulingen_US
dc.subjectAdditive manufacturingen_US
dc.subjectDynamic order arrivalen_US
dc.subjectDueling DQNen_US
dc.titleOut-of-order execution enabled deep reinforcement learning for dynamic additive manufacturing schedulingen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume91-
dc.identifier.doi10.1016/j.rcim.2024.102841-
dcterms.abstractAdditive Manufacturing (AM) has revolutionized the production landscape by enabling on-demand customized manufacturing. However, the efficient management of dynamic AM orders poses significant challenges for production planning and scheduling. This paper addresses the dynamic scheduling problem considering batch processing, random order arrival and machine eligibility constraints, aiming to minimize total tardiness in a parallel non-identical AM machine environment. To tackle this problem, we propose the out-of-order enabled dueling deep Q network (O3-DDQN) approach. In the proposed approach, the problem is formulated as a Markov decision process (MDP). Three-dimensional features, encompassing dynamic orders, AM machines, and delays, are extracted using a ‘look around’ method to represent the production status at a rescheduling point. Additionally, five novel composite scheduling rules based on the out-of-order principle are introduced for selection when an AM machine completes processing or a new order arrives. Moreover, we design a reward function that is strongly correlated with the objective to evaluate the agent's chosen action. Experimental results demonstrate the superiority of the O3-DDQN approach over single scheduling rules, randomly selected rules, and the classic DQN method. The average improvement rate of performance reaches 13.09% compared to composite scheduling rules and random rules. Additionally, the O3-DDQN outperforms the classic DQN agent with a 6.54% improvement rate. The O3-DDQN algorithm improves scheduling in dynamic AM environments, enhancing productivity and on-time delivery. This research contributes to advancing AM production and offers insights into efficient resource allocation.-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationRobotics and computer - integrated manufacturing, Feb.2025, v. 91, 102841-
dcterms.isPartOfRobotics and computer - integrated manufacturing-
dcterms.issued2025-02-
dc.identifier.scopus2-s2.0-85199899989-
dc.identifier.artn102841-
dc.description.validate202509 bcrc-
dc.description.oaNot applicableen_US
dc.identifier.FolderNumbera4084ben_US
dc.identifier.SubFormID52057en_US
dc.description.fundingSourceRGCen_US
dc.description.fundingTextNational Natural Science Foundation of China (No. 52305557); Guangdong Basic and Applied Basic Research Foundation (No. 2024A1515011930); Innovation and Technology Fund (No. PRP/038/24LI); Open Fund of State Key Laboratory of Intelligent Manufacturing Equipment and Technology (No. IMETKF2024022);en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2027-02-28en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2027-02-28
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.