Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/118091
| Title: | Egocentric human-object interaction detection : a new benchmark and method | Authors: | Deng, K Wang, Y Chau, LP |
Issue Date: | 5-Mar-2026 | Source: | Expert systems with applications, 5 Mar. 2026, v. 300, 130216 | Abstract: | Egocentric human-object interaction (Ego-HOI) detection is crucial for intelligent agents to comprehend and assist human activities from a first-person perspective. However, progress has been hindered by the lack of dedicated benchmarks and methods robust to severe egocentric challenges like hand-object occlusion. This work bridges this gap through three key contributions. Firstly, we introduce Ego-HOIBench, a pioneering benchmark dataset derived from HOI4D for real-world Ego-HOI detection, comprising over 27K real images with explicit, fine-grained <hand, verb, object> triplet annotations. Secondly, we propose Hand Geometry and Interactivity Refinement (HGIR), a novel plug-and-play module that captures the structural geometry of hands to learn occlusion-robust, pose-aware interaction representations. Thirdly, comprehensive experiments demonstrate that HGIR significantly enhances Ego-HOI detection performance across various methods, achieving state-of-the-art results and laying a solid foundation for future research in egocentric vision. Project page:https://dengkunyuan.github.io/EgoHOIBench/ | Keywords: | Egocentric vision HOI detection Human-object interaction Interaction recognition |
Publisher: | Pergamon Press | Journal: | Expert systems with applications | ISSN: | 0957-4174 | EISSN: | 1873-6793 | DOI: | 10.1016/j.eswa.2025.130216 |
| Appears in Collections: | Journal/Magazine Article |
Show full item record
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



