Seqtrack: Sequence to sequence learning for visual object tracking

X Chen, H Peng, D Wang, H Lu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
In this paper, we present a new sequence-to-sequence learning framework for visual
tracking, dubbed SeqTrack. It casts visual tracking as a sequence generation problem …

Mixformer: End-to-end tracking with iterative mixed attention

Y Cui, C Jiang, L Wang, G Wu - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Tracking often uses a multi-stage pipeline of feature extraction, target information
integration, and bounding box estimation. To simplify this pipeline and unify the process of …

Joint feature learning and relation modeling for tracking: A one-stream framework

B Ye, H Chang, B Ma, S Shan, X Chen - European Conference on …, 2022 - Springer
The current popular two-stream, two-stage tracking framework extracts the template and the
search region features separately and then performs relation modeling, thus the extracted …

Visual prompt multi-modal tracking

J Zhu, S Lai, X Chen, D Wang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Visible-modal object tracking gives rise to a series of downstream multi-modal tracking
tributaries. To inherit the powerful representations of the foundation model, a natural modus …

Aiatrack: Attention in attention for transformer visual tracking

S Gao, C Zhou, C Ma, X Wang, J Yuan - European Conference on …, 2022 - Springer
Transformer trackers have achieved impressive advancements recently, where the attention
mechanism plays an important role. However, the independent correlation computation in …

Transforming model prediction for tracking

C Mayer, M Danelljan, G Bhat, M Paul… - Proceedings of the …, 2022 - openaccess.thecvf.com
Optimization based tracking methods have been widely successful by integrating a target
model prediction module, providing effective global reasoning by minimizing an objective …

Learning spatio-temporal transformer for visual tracking

B Yan, H Peng, J Fu, D Wang… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
In this paper, we present a new tracking architecture with an encoder-decoder transformer
as the key component. The encoder models the global spatio-temporal feature …

Track anything: Segment anything meets videos

J Yang, M Gao, Z Li, S Gao, F Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Recently, the Segment Anything Model (SAM) gains lots of attention rapidly due to its
impressive segmentation performance on images. Regarding its strong ability on image …

Transformer tracking with cyclic shifting window attention

Z Song, J Yu, YPP Chen… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Transformer architecture has been showing its great strength in visual object tracking, for its
effective attention mechanism. Existing transformer-based approaches adopt the pixel-to …

RFN-Nest: An end-to-end residual fusion network for infrared and visible images

H Li, XJ Wu, J Kittler - Information Fusion, 2021 - Elsevier
In the image fusion field, the design of deep learning-based fusion methods is far from
routine. It is invariably fusion-task specific and requires a careful consideration. The most …