A Multi-Object Tracking Approach Combining Contextual Features and Trajectory Prediction
Abstract
:1. Introduction
- (1)
- During the tracking process, both the appearance and motion information of the object are taken into account.
- (2)
- Contextual features as well as individual features are used to better represent similar-looking objects.
- (3)
- The graph model uses different weights depending on the distance between objects.
- (4)
- A preprocessing module is applied to categorize the object velocities and exclude unnecessary connections between the graph model objects.
- (5)
- Post-processing of temporarily unmatched trajectories is conducted to reduce the impact of occlusion and other factors on the object tracking results.
2. Related Works
2.1. Traditional Multi-Object Tracking Methods
2.2. Multi-Object Tracking Method Based on Object Detection
3. Feature Extraction Methods
3.1. Kalman Trajectory Prediction
- The predicted value of the state in the k − 1 frame and the predetermined equations of motion are used to estimate the state of the object, and the estimated value of the state of the object is obtained.
- (1)
- Calculate the predicted values:
- (2)
- Calculate the a priori covariance matrix between the true and predicted values at the current moment:
- Correct the state estimate with the current moment’s detection result to obtain the final object state prediction.
- (3)
- Calculate the Kalman gain:
- (4)
- Estimate the true state of the object based on measured values and calculate predicted values:
- (5)
- Calculate a posteriori covariance matrix for the error between the true and estimated values:
3.2. Graph Neural Networks
3.3. Processing Methods for Contextual Features
4. Object Linkages
4.1. Hungarian Algorithm
- (1)
- First, the bipartite graph is constructed, the vertices are numbered, and the relationship between the vertices of the two sets is represented by the adjacency list.
- (2)
- Take the unmatched vertices in set M, 𝑖, and then traverse the vertices connected to it in set N. If the vertices in set N do not match, then match the two points. Then, continue to traverse the unmatched vertices in set M using the same principle. If the vertex in set N is already matched, then try to recursively match the matching point of the matched vertex in set N with another vertex match.
- (3)
- Perform step (2) recursively until all vertices in M have been traversed to obtain the maximum match [29].
4.2. Trajectory Post-Processing
- When N = 0, the similarity matrix is first calculated based on the contextual features. Then, the similarity matrix obtained from the Mahalanobis distance, as well as the intersection and merger ratio, is used as constraints to exclude matching errors due to occlusion, etc. This stage completes the majority of the matches.
- When N > 0, the contextual features of neighboring objects become more distinguishable, and we obtain the similarity matrix for matching based on these contextual features.
- When N ≥ 3, we need to match the object in two frames that are far apart. Due to the object’s movement, the contextual features of the object tend to change significantly. Therefore, in such cases, we match the object based on its individual features.
5. Analysis of Experimental Results
5.1. Assessment Criteria
5.2. Simulation Analysis
6. Summary
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Blackman, S.S. Multiple hypothesis tracking for multiple object tracking. IEEE Aerosp. Electron. Syst. Mag. 2004, 19, 5–18. [Google Scholar] [CrossRef]
- Dendorfer, P.; Osep, A.; Milan, A.; Schindler, K.; Cremers, D.; Reid, I.; Leal-Taixé, L. Motchallenge: A benchmak for single-camera multiple object tracking. Int. J. Comput. Vis. 2021, 129, 845–881. [Google Scholar] [CrossRef]
- Zheng, L.; Tang, M.; Chen, Y.; Zhu, G.; Wang, J.; Lu, H. Improving multiple object tracking with single object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Beijing, China, 19–25 June 2021; pp. 2453–2462. [Google Scholar]
- Wang, Q.; Zheng, Y.; Pan, P.; Xu, Y. Multiple object tracking with correlation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Beijing, China, 19–25 June 2021; pp. 3876–3886. [Google Scholar]
- Liu, Z.; Shang, Y.; Li, T.; Chen, G.; Wang, Y.; Hu, Q.; Zhu, P. Robust Multi-Drone Multi-object Tracking to Resolve object Occlusion: A Benchmark. IEEE Trans. Multimed. 2023, 25, 1462–1476. [Google Scholar] [CrossRef]
- Yang, D. Research on multi-object tracking technology based on machine vision. Appl. Nanosci. 2023, 13, 2945–2955. [Google Scholar] [CrossRef]
- Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
- Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
- Sun, P.; Cao, J.; Jiang, Y. Transtrack: Multiple object tracking with transformer. arXiv 2020, arXiv:2012.15460. [Google Scholar]
- Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. Fairmot: On the fairness of detection and re-identification in multiple object tracking. Int. J. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
- Wang, Z.; Zheng, L.; Liu, Y.; Li, Y.; Wang, S. Towards real-time multi-object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 3069–3087. [Google Scholar]
- Aharon, N.; Orfaig, R.; Bobrovsky, B.Z. BoT-SORT: Robust associations multi-pedestrian tracking. arXiv 2022, arXiv:2206.14651. [Google Scholar]
- Azhar, M.I.H.; Zaman, F.H.K.; Tahir, N.M.; Hashim, H. People tracking system using DeepSORT. In Proceedings of the 2020 10th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 21–22 August 2020; pp. 137–141. [Google Scholar]
- Wang, Y.H. SMILEtrack: SiMIlarity LEarning for Multiple Object Tracking. arXiv 2022, arXiv:2211.08824. [Google Scholar]
- Voigtlaender, P.; Krause, M.; Osep, A.; Luiten, J.; Sekar, B.B.G.; Geiger, A.; Leibe, B. Mots: Multi-object tracking and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 7942–7951. [Google Scholar]
- Dendorfer, P.; Rezatofighi, H.; Milan, A.; Shi, J.; Cremers, D.; Reid, I.; Leal-Taixé, L. Mot20: A benchmark for multi object tracking in crowded scenes. arXiv 2020, arXiv:2003.09003. [Google Scholar]
- Xu, T.; Feng, Z.H.; Wu, X.J.; Kittler, J. Learning Adaptive Discriminative Correlation Filters via Temporal Consistency Preserving Spatial Feature Selection for Robust Visual Object Tracking. IEEE Trans. Image Process. 2019, 28, 5596–5609. [Google Scholar] [CrossRef] [PubMed]
- Wang, Y.; Kitani, K.; Weng, X. Joint object detection and multi-object tracking with graph neural networks. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13708–13715. [Google Scholar]
- Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual Object Tracking Using Adaptiv--**e Correlation Filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
- Liu, F.; Gong, C.; Huang, X.; Zhou, T.; Yang, J.; Tao, D. Robust Visual Tracking Revisited: From Correlation Filter to Template Matching. IEEE Trans. Image Process. 2018, 27, 2777–2790. [Google Scholar] [CrossRef] [PubMed]
- Wen, L.; Du, D.; Cai, Z.; Lei, Z.; Chang, M.; Qi, H.; Lyu, S. UA-DETRAC: A new benchmark and protocol for multi-object detection and tracking. Comput. Vis. Image Underst. 2020, 193, 102907. [Google Scholar] [CrossRef]
- Zeng, F.; Dong, B.; Wang, T.; Chen, C.; Zhang, X.; Wei, Y. Motr: End-to-end multiple-object tracking with transformer. arXiv 2021, arXiv:2105.03247. [Google Scholar]
- Zhang, W.; Zhou, H.; Sun, S.; Wang, Z.; Shi, J.; Loy, C.C. Robust multi-modality multi-object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2365–2374. [Google Scholar]
- Luiten, J.; Osep, A.; Dendorfer, P.; Torr, P.; Geiger, A.; Leal-Taixé, L.; Leibe, B. Hota: A higher order metric for evaluating multi-object tracking. Int. J. Comput. Vis. 2021, 129, 548–578. [Google Scholar] [CrossRef]
- Sulikowski, P.; Zdziebko, T. Deep Learning-Enhanced Framework for Performance Evaluation of a Recommending Interface with Varied Recommendation Position and Intensity Based on Eye-Tracking Equipment Data Processing. Electronics 2020, 9, 266. [Google Scholar] [CrossRef]
- Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Wang, X. Bytetrack: Multi-object tracking by associating every detection box. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 1–21. [Google Scholar]
- Jiang, X.; Ma, T.; Jin, J.; Jiang, Y. Sensor Management with Dynamic Clustering for Bearings-Only Multi-object Tracking via Swarm Intelligence Optimization. Electronics 2023, 12, 3397. [Google Scholar] [CrossRef]
- Cheng, Y.; Zhang, S.; Wang, X.; Wang, H. Self-Tuning Process Noise in Variational Bayesian Adaptive Kalman Filter for object Tracking. Electronics 2023, 12, 3887. [Google Scholar] [CrossRef]
- Zheng, X.; Liu, Y.; Pan, S.; Zhang, M.; Jin, D.; Yu, P.S. Graph neural networks for graphs with heterophily: A survey. arXiv 2022, arXiv:2202.07082. [Google Scholar]
- Chu, P.; Fan, H.; Tan, C.C.; Ling, H. Online multi-object tracking with instance-aware tracker and dynamic model refreshment. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 7–11 January 2019; pp. 161–170. [Google Scholar]
- Chen, Z.; Xiong, X.; Meng, F.; Xiao, X.; Liu, J. Scaling-Invariant Max-Filtering Enhancement Transformers for Efficient Visual Tracking. Electronics 2023, 12, 3905. [Google Scholar] [CrossRef]
Algorithm | MOTA () | IDF1 () | IDS () | FPS (Frame/s) |
---|---|---|---|---|
Transcenter | 73.2 | 62.4 | 4614 | 1 |
Transtrack | 75.2 | 63.5 | 3603 | 59.2 |
Fairmot | 73.7 | 72.3 | 3303 | 25.9 |
SUSHI | 81 | 83 | 1149 | 21 |
Bytetrack | 78.9 | 77.2 | 2359 | 29.8 |
Proposed | 80 | 84 | 2303 | 55 |
Similarity Type | Scene 1 | Scene 2 | ||
---|---|---|---|---|
MOTA () | IDS () | MOTA () | IDS () | |
Similarity of motion features | 69.7 | 96 | 72.6 | 214 |
Similarity of appearance features | 72.4 | 95 | 74.7 | 159 |
Motion + Appearance features similarity | 84.2 | 90 | 80.5 | 103 |
Motion + Appearance + Preprocessing similarity | 88.9 | 72 | 82.6 | 99 |
Motion + Appearance + PostProcessing Similarity | 94.3 | 68 | 86 | 92 |
Proposed Similarity | 96.1 | 6 | 92 | 20 |
Real similarity | 99.3 | 1 | 92.4 | 4 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, P.; Jing, Q.; Zhao, X.; Dong, L.; Lei, W.; Zhang, W.; Lin, Z. A Multi-Object Tracking Approach Combining Contextual Features and Trajectory Prediction. Electronics 2023, 12, 4720. https://doi.org/10.3390/electronics12234720
Zhang P, Jing Q, Zhao X, Dong L, Lei W, Zhang W, Lin Z. A Multi-Object Tracking Approach Combining Contextual Features and Trajectory Prediction. Electronics. 2023; 12(23):4720. https://doi.org/10.3390/electronics12234720
Chicago/Turabian StyleZhang, Peng, Qingyang Jing, Xinlei Zhao, Lijia Dong, Weimin Lei, Wei Zhang, and Zhaonan Lin. 2023. "A Multi-Object Tracking Approach Combining Contextual Features and Trajectory Prediction" Electronics 12, no. 23: 4720. https://doi.org/10.3390/electronics12234720
APA StyleZhang, P., Jing, Q., Zhao, X., Dong, L., Lei, W., Zhang, W., & Lin, Z. (2023). A Multi-Object Tracking Approach Combining Contextual Features and Trajectory Prediction. Electronics, 12(23), 4720. https://doi.org/10.3390/electronics12234720