Learning Dual-Fused Modality-Aware Representations for RGBD Tracking
With the development of depth sensors in recent years, RGBD object tracking has received significant attention. Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference. However, some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored. On the other hand, some methods attempt to fuse the two modalities by treating them equally, resulting in the missing of modality-specific features. To tackle these limitations, we propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking. The first fusion module focuses on extracting the shared information between modalities based on cross-modal attention. The second aims at integrating the RGB-specific and depth-specific information to enhance the fused features. By fusing both the modality-shared and modality-specific information in a modality-aware scheme, our DMTracker can learn discriminative representations in complex tracking scenes. Experiments show that our proposed tracker achieves very promising results on challenging RGBD benchmarks. Code is available at https://github.com/ShangGaoG/DMTracker.
First ; Corresponding
Cited Times [WOS]:0
|Document Type||Conference paper|
|Department||Department of Computer Science and Engineering|
1.Department of Computer Science and Engineering,Southern University of Science and Technology,Shenzhen,China
2.University of Birmingham,Birmingham,United Kingdom
3.University of Electronic Science and Technology of China,Chengdu,China
|First Author Affilication||Department of Computer Science and Engineering|
|Corresponding Author Affilication||Department of Computer Science and Engineering|
|First Author's First Affilication||Department of Computer Science and Engineering|
Gao，Shang,Yang，Jinyu,Li，Zhe,et al. Learning Dual-Fused Modality-Aware Representations for RGBD Tracking[C],2023:478-494.
|Files in This Item:||There are no files associated with this item.|
|Recommend this item|
|Export to Endnote|
|Export to Excel|
|Export to Csv|
|Similar articles in Google Scholar|
|Similar articles in Baidu Scholar|
|Similar articles in Bing Scholar|
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.