Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/82194
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electronic and Information Engineering-
dc.creatorLi, HJ-
dc.creatorWu, SH-
dc.creatorHuang, SP-
dc.creatorLam, KM-
dc.creatorXing, XF-
dc.date.accessioned2020-05-05T05:59:03Z-
dc.date.available2020-05-05T05:59:03Z-
dc.identifier.urihttp://hdl.handle.net/10397/82194-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rightsThis work is licensed under a Creative Commons Attribution 4.0 License. For more information, see http://creativecommons.org/licenses/by/4.0/en_US
dc.rightsThe following publication H. Li, S. Wu, S. Huang, K. Lam and X. Xing, "Deep Motion-Appearance Convolutions for Robust Visual Tracking," in IEEE Access, vol. 7, pp. 180451-180466, 2019 is available at https://dx.doi.org/10.1109/ACCESS.2019.2958405en_US
dc.subjectVisual trackingen_US
dc.subject3D convolutional kernelsen_US
dc.subjectMotion-appearanceen_US
dc.titleDeep motion-appearance convolutions for robust visual trackingen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage180451-
dc.identifier.epage180466-
dc.identifier.volume7-
dc.identifier.doi10.1109/ACCESS.2019.2958405-
dcterms.abstractVisual tracking is a challenging task due to unconstrained appearance variations and dynamic surrounding backgrounds, which basically arise from the complex motion of the target object. Therefore, the information and the correlation between the target motion and its resulting appearance should be considered comprehensively to achieve robust tracking performance. In this paper, we propose a deep neural network for visual tracking, namely the Motion-Appearance Dual (MADual) network, which employs a dual-branch architecture, by using deep two-dimensional (2D) and deep three-dimensional (3D) convolutions to integrate the local and global information of the target object's motion and appearance synchronously. For each frame of a tracking video, 2D convolutional kernels of the deep 2D branch slide over the frame to extract its global spatial-appearance features. Meanwhile, 3D convolutional kernels of the deep 3D branch are used to collaboratively extract the appearance and the associated motion features of the visual target from successive frames. By sliding the 3D convolutional kernels along a video sequence, the model is able to learn the temporal features from previous frames, and therefore, generate the local patch-based motion patterns of the target. Sliding the 2D kernels on a frame and the 3D kernels on a frame cube synchronously enables a better hierarchical motion-appearance integration, and boosts the performance for the visual tracking task. To further improve the tracking precision, an extra ridge-regression model is trained for the tracking process, based not only on the bounding box given in the first frame, but also on its synchro-frame-cube using our proposed Inverse Temporal Training method (ITT). Extensive experiments on popular benchmark datasets, OTB2013, OTB50, OTB2015, UAV123, TC128, VOT2015 and VOT2016, demonstrate that the proposed MADual tracker performs favorably against many state-of-the-art methods.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIEEE access, 9 Dec. 2019, v. 7, p. 180451-180466-
dcterms.isPartOfIEEE access-
dcterms.issued2019-
dc.identifier.isiWOS:000509483800250-
dc.identifier.scopus2-s2.0-85077215831-
dc.identifier.eissn2169-3536-
dc.description.validate202006 bcrc-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.pubStatusPublisheden_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Li_Motion-Appearance_Convolutions_Visual.pdf2.14 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

75
Last Week
1
Last month
Citations as of Apr 21, 2024

Downloads

66
Citations as of Apr 21, 2024

SCOPUSTM   
Citations

2
Citations as of Apr 19, 2024

WEB OF SCIENCETM
Citations

1
Citations as of Apr 18, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.