TY - JOUR
T1 - Learning modality-consistency feature templates
T2 - A robust RGB-infrared tracking system
AU - Lan, Xiangyuan
AU - Ye, Mang
AU - Shao, Rui
AU - Zhong, Bineng
AU - Yuen, Pong Chi
AU - Zhou, Huiyu
N1 - Funding Information:
Manuscript received September 9, 2018; revised November 16, 2018 and January 12, 2019; accepted January 14, 2019. Date of publication February 15, 2019; date of current version July 31, 2019. This work was supported in part by the Hong Kong Research Grants Council RGC/HKBU12254316 and by the Hong Kong Baptist University Tier 1 Start-up Grant. The work of H. Zhou was supported by the U.K. Engineering and Physical Sciences Research Council under Grant EP/N011074/1, by the Royal Society-Newton Advanced Fellowship under Grant NA160342, and by the European Union’s Horizon 2020 research and innovation program under the Marie-Sklodowska-Curie Grant 720325. The work of B. Zhong was supported by the National Natural Science Foundation of China under Grant 61572205. (Corresponding author: Pong C. Yuen.) X. Lan, M. Ye, R. Shao, and P. C. Yuen are with the Department of Computer Science, Hong Kong Baptist University, Hong Kong SAR, China (e-mail:, [email protected]; [email protected]. edu.hk; [email protected]; [email protected]).
PY - 2019/12
Y1 - 2019/12
N2 - With a large number of video surveillance systems installed for the requirement from industrial security, the task of object tracking, which aims to locate objects of interest in videos, is very important. Although numerous tracking algorithms for RGB videos have been developed in the decade, the tracking performance and robustness of these systems may be degraded dramatically when the information from RGB video is unreliable (e.g., poor illumination conditions or very low resolution). To address this issue, this paper presents a new tracking system, which aims to combine the information from RGB and infrared modalities for object tracking. The proposed tracking systems is based on our proposed machine learning model. Particularly, the learning model can alleviate the modality discrepancy issue under the proposed modality consistency constraint from both representation patterns and discriminability, and generate discriminative feature templates for collaborative representations and discrimination in heterogeneous modalities. Experiments on a variety of challenging RGB-infrared videos demonstrate the effectiveness of the proposed algorithm.
AB - With a large number of video surveillance systems installed for the requirement from industrial security, the task of object tracking, which aims to locate objects of interest in videos, is very important. Although numerous tracking algorithms for RGB videos have been developed in the decade, the tracking performance and robustness of these systems may be degraded dramatically when the information from RGB video is unreliable (e.g., poor illumination conditions or very low resolution). To address this issue, this paper presents a new tracking system, which aims to combine the information from RGB and infrared modalities for object tracking. The proposed tracking systems is based on our proposed machine learning model. Particularly, the learning model can alleviate the modality discrepancy issue under the proposed modality consistency constraint from both representation patterns and discriminability, and generate discriminative feature templates for collaborative representations and discrimination in heterogeneous modalities. Experiments on a variety of challenging RGB-infrared videos demonstrate the effectiveness of the proposed algorithm.
KW - Multimodal sensor fusion
KW - tracking system
KW - video surveillance system
UR - http://www.scopus.com/inward/record.url?scp=85070494127&partnerID=8YFLogxK
U2 - 10.1109/TIE.2019.2898618
DO - 10.1109/TIE.2019.2898618
M3 - Journal article
AN - SCOPUS:85070494127
SN - 0278-0046
VL - 66
SP - 9887
EP - 9897
JO - IEEE Transactions on Industrial Electronics
JF - IEEE Transactions on Industrial Electronics
IS - 12
ER -