Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116348
DC FieldValueLanguage
dc.contributorDepartment of Industrial and Systems Engineeringen_US
dc.creatorWu, Den_US
dc.creatorZhao, Qen_US
dc.creatorFan, Jen_US
dc.creatorQi, Jen_US
dc.creatorZheng, Pen_US
dc.creatorHu, Jen_US
dc.date.accessioned2025-12-18T06:39:42Z-
dc.date.available2025-12-18T06:39:42Z-
dc.identifier.issn0278-6125en_US
dc.identifier.urihttp://hdl.handle.net/10397/116348-
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.subjectFew-shot learningen_US
dc.subjectHuman–robot collaborationen_US
dc.subjectIntent recognitionen_US
dc.subjectVision-language modelsen_US
dc.titleH2R bridge : transferring vision-language models to few-shot intention meta-perception in human robot collaborationen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage524en_US
dc.identifier.epage535en_US
dc.identifier.volume80en_US
dc.identifier.doi10.1016/j.jmsy.2025.03.016en_US
dcterms.abstractHuman–robot collaboration enhances efficiency by enabling robots to work alongside human operators in shared tasks. Accurately understanding human intentions is critical for achieving a high level of collaboration. Existing methods heavily rely on case-specific data and face challenges with new tasks and unseen categories, while often limited data is available under real-world conditions. To bolster the proactive cognitive abilities of collaborative robots, this work introduces a Visual-Language-Temporal approach, conceptualizing intent recognition as a multimodal learning problem with HRC-oriented prompts. A large model with prior knowledge is fine-tuned to acquire industrial domain expertise, then enables efficient rapid transfer through few-shot learning in data-scarce scenarios. Comparisons with state-of-the-art methods across various datasets demonstrate the proposed approach achieves new benchmarks. Ablation studies confirm the efficacy of the multimodal framework, and few-shot experiments further underscore meta-perceptual potential. This work addresses the challenges of perceptual data and training costs, building a human–robot bridge (H2R Bridge) for semantic communication, and is expected to facilitate proactive HRC and further integration of large models in industrial applications.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationJournal of manufacturing systems, June 2025, v. 80, p. 524-535en_US
dcterms.isPartOfJournal of manufacturing systemsen_US
dcterms.issued2025-06-
dc.identifier.scopus2-s2.0-105001851845-
dc.description.validate202512 bchyen_US
dc.description.oaNot applicableen_US
dc.identifier.SubFormIDG000494/2025-12-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work is supported by the National Natural Science Foundation of China (Grant Nos. U23B20102 , 52475270 , 52375254 ) and Xie Youbai Design Scientific Research Foundation ( XYB-DS-202401 ).en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2027-06-30en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2027-06-30
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

10
Citations as of Apr 3, 2026

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.