Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/106607
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Mechanical Engineering | en_US |
| dc.creator | Zhang, S | en_US |
| dc.creator | Zhang, D | en_US |
| dc.creator | Zou, Q | en_US |
| dc.date.accessioned | 2024-05-14T05:42:05Z | - |
| dc.date.available | 2024-05-14T05:42:05Z | - |
| dc.identifier.issn | 1380-7501 | en_US |
| dc.identifier.uri | http://hdl.handle.net/10397/106607 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Springer | en_US |
| dc.rights | © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024 | en_US |
| dc.rights | This version of the article has been accepted for publication, after peer review (when applicable) and is subject to Springer Nature’s AM terms of use (https://www.springernature.com/gp/open-research/policies/accepted-manuscript-terms), but is not the Version of Record and does not reflect post-acceptance improvements, or any corrections. The Version of Record is available online at: http://dx.doi.org/10.1007/s11042-024-19002-4. | en_US |
| dc.subject | Visual object tracking | en_US |
| dc.subject | Global-local representation aggregation | en_US |
| dc.subject | Channel information | en_US |
| dc.subject | Transformer attention | en_US |
| dc.subject | Convolution | en_US |
| dc.title | TGLC : visual object tracking by fusion of global-local information and channel information | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.spage | 89151 | en_US |
| dc.identifier.epage | 89172 | en_US |
| dc.identifier.volume | 83 | en_US |
| dc.identifier.issue | 41 | en_US |
| dc.identifier.doi | 10.1007/s11042-024-19002-4 | en_US |
| dcterms.abstract | Visual object tracking aspires to locate the target incessantly in each frame with designated initial target location, which is an imperative yet demanding task in computer vision. Recent approaches strive to fuse global information of template and search region for object tracking, which achieve promising tracking performance. However, fusion of global information devastates some local details. Local information is essential for distinguishing the target from background regions. With a focus on addressing this problem, this work presents a novel tracking algorithm TGLC integrating a channel-aware convolution block and Transformer attention for global and local representation aggregation, and for channel information modeling. This method is capable of accurately estimating the bounding box of the target. Extensive experiments are conducted on five widely recognized datasets, i.e., GOT-10k, TrackingNet, LaSOT, OTB100 and UAV123. The results depict that the proposed tracking method achieves competitive tracking performance compared with state-of-the-art trackers while still running in real-time. Visualization of the tracking results on LaSOT further demonstrates the capability of the proposed tracking method to cope with tracking challenges, e.g., illumination variation, deformation of the target and background clutter. | en_US |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | Multimedia tools and applications, Dec. 2024, v. 83, no. 41, p. 89151-89172 | en_US |
| dcterms.isPartOf | Multimedia tools and applications | en_US |
| dcterms.issued | 2024-12 | - |
| dc.description.validate | 202405 bcrc | en_US |
| dc.description.oa | Accepted Manuscript | en_US |
| dc.identifier.FolderNumber | a2698 | - |
| dc.identifier.SubFormID | 48069 | - |
| dc.description.fundingSource | Self-funded | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Green (AAM) | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Zhang_TGLC_Visual_Object.pdf | 1.74 MB | Adobe PDF | View/Open |
Page views
47
Citations as of Apr 14, 2025
Downloads
5
Citations as of Apr 14, 2025
Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



