Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/112643
DC Field | Value | Language |
---|---|---|
dc.contributor | Department of Electrical and Electronic Engineering | - |
dc.creator | Lu, CK | - |
dc.creator | Mak, MW | - |
dc.creator | Li, RM | - |
dc.creator | Chi, ZR | - |
dc.creator | Fu, H | - |
dc.date.accessioned | 2025-04-24T00:28:17Z | - |
dc.date.available | 2025-04-24T00:28:17Z | - |
dc.identifier.uri | http://hdl.handle.net/10397/112643 | - |
dc.language.iso | en | en_US |
dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
dc.rights | © 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. | en_US |
dc.rights | For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/ | en_US |
dc.rights | The following publication C. -K. Lu, M. -W. Mak, R. Li, Z. Chi and H. Fu, "Action Progression Networks for Temporal Action Detection in Videos," in IEEE Access, vol. 12, pp. 126829-126844, 2024 is available at https://dx.doi.org/10.1109/ACCESS.2024.3451503. | en_US |
dc.subject | Action recognition | en_US |
dc.subject | Temporal action detection | en_US |
dc.subject | Video analysis | en_US |
dc.title | Action progression networks for temporal action detection in videos | en_US |
dc.type | Journal/Magazine Article | en_US |
dc.identifier.spage | 126829 | - |
dc.identifier.epage | 126844 | - |
dc.identifier.volume | 12 | - |
dc.identifier.doi | 10.1109/ACCESS.2024.3451503 | - |
dcterms.abstract | This study introduces an innovative Temporal Action Detection (TAD) model that is distinguished by its lightweight structure and capability for end-to-end training, delivering competitive performance. Traditional TAD approaches often rely on pre-trained models for feature extraction, compromising on end-to-end training for efficiency, yet encounter challenges due to misalignment with tasks and data shifts. Our method addresses these challenges by processing untrimmed videos on a snippet basis, facilitating a snippet-level TAD model that is trained end-to-end. Central to our approach is a novel frame-level label, termed action progressions, designed to encode temporal localization information. The prediction of action progressions not only enables our snippet-level model to incorporate temporal information effectively but also introduces a granular temporal encoding for the evolution of actions, enhancing the precision of detection. Beyond a streamlined pipeline, our model introduces several novel capabilities: 1) It directly learns from raw videos, unlike prevalent TAD methods that depend on frozen, pre-trained feature extraction models; 2) It is flexible for training with trimmed and untrimmed videos; 3) It is the first TAD model to avoid the detection of incomplete actions; and 4) It can accurately detect long-lasting actions or those with clear evolutionary patterns. Utilizing these advantages, our model achieves commendable performance on benchmark datasets, securing averaged mean Average Precision (mAP) scores of 54.8%, 30.5%, and 78.7% on THUMOS14, ActivityNet-1.3, and DFMAD, respectively. | - |
dcterms.accessRights | open access | en_US |
dcterms.bibliographicCitation | IEEE access, 2024, v. 12, p. 126829-126844 | - |
dcterms.isPartOf | IEEE access | - |
dcterms.issued | 2024 | - |
dc.identifier.isi | WOS:001316135100001 | - |
dc.identifier.eissn | 2169-3536 | - |
dc.description.validate | 202504 bcrc | - |
dc.description.oa | Version of Record | en_US |
dc.identifier.FolderNumber | OA_Scopus/WOS | en_US |
dc.description.fundingSource | RGC | en_US |
dc.description.fundingSource | Others | en_US |
dc.description.fundingText | Hong Kong Polytechnic University | en_US |
dc.description.pubStatus | Published | en_US |
dc.description.oaCategory | CC | en_US |
Appears in Collections: | Journal/Magazine Article |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Lu_Action_Progression_Networks.pdf | 1.93 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.