Journal Home > Volume 6 , Issue 4

In autonomous driving, target tracking is essential to environmental perception. The study of target tracking algorithms can improve the accuracy of an autonomous driving vehicle’s perception, which is of great significance in ensuring the safety of autonomous driving and promoting the landing of technical applications. This study focuses on the fusion tracking algorithm based on visible and infrared images. The proposed approach utilizes a feature-level image fusion method, dividing the tracking process into two components: image fusion and target tracking. An unsupervised network, Visible and Infrared image Fusion Network (VIF-net), is employed for visible and infrared image fusion in the image fusion part. In the target tracking part, Siamese Region Proposal Network (SiamRPN), based on deep learning, tracks the target with fused images. The fusion tracking algorithm is trained and evaluated on the visible infrared image dataset RGBT234. Experimental results demonstrate that the algorithm outperforms training networks solely based on visible images, proving that the fusion of visible and infrared images in the target tracking algorithm can improve the accuracy of the target tracking even if it is like tracking-based visual images. This improvement is also attributed to the algorithm’s ability to extract infrared image features, augmenting the target tracking accuracy.


menu
Abstract
Full text
Outline
About this article

Enhanced target tracking algorithm for autonomous driving based on visible and infrared image fusion

Show Author's information Quan Yuan( )Haixu ShiAshton Tan Yu XuanMing GaoQing XuJianqiang Wang
State Key Laboratory of Intelligent Green Vehicle and Mobility, School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China

Abstract

In autonomous driving, target tracking is essential to environmental perception. The study of target tracking algorithms can improve the accuracy of an autonomous driving vehicle’s perception, which is of great significance in ensuring the safety of autonomous driving and promoting the landing of technical applications. This study focuses on the fusion tracking algorithm based on visible and infrared images. The proposed approach utilizes a feature-level image fusion method, dividing the tracking process into two components: image fusion and target tracking. An unsupervised network, Visible and Infrared image Fusion Network (VIF-net), is employed for visible and infrared image fusion in the image fusion part. In the target tracking part, Siamese Region Proposal Network (SiamRPN), based on deep learning, tracks the target with fused images. The fusion tracking algorithm is trained and evaluated on the visible infrared image dataset RGBT234. Experimental results demonstrate that the algorithm outperforms training networks solely based on visible images, proving that the fusion of visible and infrared images in the target tracking algorithm can improve the accuracy of the target tracking even if it is like tracking-based visual images. This improvement is also attributed to the algorithm’s ability to extract infrared image features, augmenting the target tracking accuracy.

Keywords: deep learning, autonomous driving, image fusion, target tracking, infrared image

References(19)

[1]
Bertinetto, L., Valmadre, J., Henriques, J. F., Vedaldi, A., Torr, P. H. S., 2016. Fully-convolutional Siamese networks for object tracking. In: European Conference on Computer Vision, 850–865.
DOI
[2]
Bolme, D., Beveridge, J. R., Draper, B. A., Lui, Y. M., 2010. Visual object tracking using adaptive correlation filters. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2544–2550.
DOI
[3]
[4]
He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770–778.
DOI
[5]
[6]
[7]
Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K. Q., 2017. Densely connected convolutional networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 4700–4708.
DOI
[8]
[9]
Kristan, M., Leonardis, A., Matas, J., Felsberg, M., Pflugfelder, R., Zajc, L., et al., 2019. The sixth visual object tracking VOT2018 challenge results. In: European Conference on Computer Vision, 3–53.
[10]
Krizhevsky, A., Sutskever, I., Hinton, G. E., 2012. ImageNet classification with deep convolutional neural networks. In: Proceedings of the 25th International Conference on Neural Information Processing Systems, 1097–1105.
[11]
Li, B., Yan, J., Wu, W., Zhu, Z., Hu, X., 2018. High performance visual tracking with Siamese region proposal network. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8971–8980.
DOI
[12]
Li, B., Wu, W., Wang, Q., Zhang, F., Xing, J., Yan, J., 2019. SiamRPN++: evolution of Siamese visual tracking with very deep networks. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 4277–4286.
DOI
[13]
[14]
[15]
[16]
Wang, Q., Zhang, L., Bertinetto, L., Hu, W., Torr, P. H. S., 2019. Fast online object tracking and segmentation: A unifying approach. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 1328–1338.
DOI
[17]
[18]
[19]
Zhu, Z., Wang, Q., Li, B., Wu, W., Yan, J., Hu, W., 2018. Distractor-aware Siamese networks for visual object tracking. In: European Conference on Computer Vision, 103–119.
DOI
Publication history
Copyright
Acknowledgements
Rights and permissions

Publication history

Received: 08 July 2023
Revised: 20 August 2023
Accepted: 12 September 2023
Published: 30 December 2023
Issue date: December 2023

Copyright

© The author(s) 2023.

Acknowledgements

Acknowledgements

The National Natural Science Foundation of China funded this research (Grant Nos. 52072214 and 52242213). The authors acknowledge Dr. Hui Xiong for his assistance in improving the manuscript.

Rights and permissions

This is an open access article under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).

Return