TensorTrack: Tensor Decomposition for Video Object Tracking
Yuntao Gu,
Pengfei Zhao,
Lan Cheng,
Yuanjun Guo (),
Haikuan Wang,
Wenjun Ding and
Yu Liu
Additional contact information
Yuntao Gu: College of Electrical and Power Engineering, Taiyuan University of Technology, Taiyuan 030024, China
Pengfei Zhao: School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
Lan Cheng: College of Electrical and Power Engineering, Taiyuan University of Technology, Taiyuan 030024, China
Yuanjun Guo: Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences Shenzhen, Shenzhen 518055, China
Haikuan Wang: School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
Wenjun Ding: China Construction Third Bureau Digital Engineering Co., Ltd., Shenzhen 518106, China
Yu Liu: China Construction Third Bureau Digital Engineering Co., Ltd., Shenzhen 518106, China
Mathematics, 2025, vol. 13, issue 4, 1-17
Abstract:
Video Object Tracking (VOT) is a critical task in computer vision. While Siamese-based and Transformer-based trackers are widely used in VOT, they struggle to perform well on the OTB100 benchmark due to the lack of dedicated training sets. This challenge highlights the difficulty of effectively generalizing to unknown data. To address this issue, this paper proposes an innovative method that utilizes tensor decomposition, an underexplored concept in object-tracking research. By applying L1-norm tensor decomposition, video sequences are represented as four-mode tensors, and a real-time background subtraction algorithm is introduced, allowing for effective modeling of the target–background relationship and adaptation to environmental changes, leading to accurate and robust tracking. Additionally, the paper integrates an improved multi-kernel correlation filter into a single frame, locating and tracking the target by comparing the correlation between the target template and the input image. To further enhance localization precision and robustness, the paper also incorporates Tucker2 decomposition to integrate appearance and motion patterns, generating composite heatmaps. The method is evaluated on the OTB100 benchmark dataset, showing significant improvements in both performance and speed compared to traditional methods. Experimental results demonstrate that the proposed method achieves a 15.8% improvement in AUC and a ten-fold increase in speed compared to typical deep learning-based methods, providing an efficient and accurate real-time tracking solution, particularly in scenarios with similar target–background characteristics, high-speed motion, and limited target movement.
Keywords: video object tracking; correlation filter; tensor decomposition; background subtraction algorithm (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/13/4/568/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/4/568/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:4:p:568-:d:1586904
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().