RACFME: Object Tracking in Satellite Videos by Rotation Adaptive Correlation Filters with Motion Estimations
Abstract
:1. Introduction
- We introduce a rotation-adaptive (RA) feature enhancement module to address the issue of the lack of rotation invariance in HOG. By using a few affine transformations, the number of effective positive samples is increased and the rotating target tracking ability of the KCF method is enhanced.
- Based on the correlation between peak response and occlusion, an occlusion detection method for vehicles and ships in satellite video is proposed. Then Kalman filter is combined with motion path equalization to predict a target when occlusion is detected, which solves the problem of target loss when partial or complete occlusion occurs.
- Our algorithm achieves a success score of 95% in tracking objects, operating at about 116 FPS, showcasing a superior performance. Moreover, the method effectively addresses tracking failures in scenarios involving occluded and rotating objects.
2. Related Work
2.1. Object Tracking in Satellite Videos
2.2. Moving Object Tracking
2.3. Kernelized Correlation Filter
3. Proposed Method
3.1. Rotation-Adaptive Feature Enhancement Module
Algorithm 1 RA module tracking |
Input: frames: video stream; Output: : position at subsequent frames; for i = 1; i < length(frames); i + + do if i == 1 then /*Select the object to track and do some initialization*/ ; R ← R0; else then ; ← Rotate z based on R; rotation angle ← Equation (6); ← The response map; ; R ← [0, rotation angle, closest angle]; return end else end for |
3.2. Motion Estimation
4. Experiment and Analysis
4.1. Datasets and Evaluation Metrics
- Small object size and limited spatial resolution: The objects in our dataset, such as vehicles and ships, are relatively small (e.g., 15 × 8 pixels for vehicles and 50 × 15 pixels for ships), and their spatial resolution is only about 1.5 m. This makes it difficult to distinguish the objects from the background, especially when the objects are partially occluded or have similar colors to the surroundings.
- Frequent occlusions: The dataset contains numerous instances of object occlusions, especially in the traffic scenarios in Hong Kong, Shanghai, and New York. Vehicles are often occluded by other vehicles or infrastructure such as bridges, which poses a significant challenge for tracking algorithms.
- Rotating objects: The dataset includes objects that undergo significant rotations, such as the rotating aircraft in the airport videos. Traditional tracking algorithms that rely on non-rotation-invariant features often fail to track these objects accurately.
- Diverse object types and motion patterns: The dataset includes different types of objects (vehicles, ships, and aircraft) with varying sizes, shapes, and motion patterns. This diversity requires the tracking algorithm to be robust to different object characteristics and motion dynamics.
4.2. Experimental Settings
4.3. Experimental Analysis on Moving Object Tracking
4.4. Experimental Analysis on Rotating Object Tracking
4.5. Experimental Analysis on Occluded Object Tracking
5. Qualitative Evaluation
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Zhang, Z.; Peng, T.; Liao, L.; Xiao, J.; Wang, M. SDM-Car: A Dataset for Small and Dim Moving Vehicles Detection in Satellite Videos. IEEE Geosci. Remote Sens. Lett. 2024, 21, 1–5. [Google Scholar] [CrossRef]
- Shangguan, D.; Chen, L.; Ding, J. A Digital Twin-Based Approach for the Fault Diagnosis and Health Monitoring of a Complex Satellite System. Symmetry 2020, 12, 1307. [Google Scholar] [CrossRef]
- Ghaban, W.; Ahmad, J.; Siddique, A.A.; Alshehri, M.S.; Saghir, A.; Saeed, F.; Ghaleb, B.; Rehman, M.U. Sustainable Environmental Monitoring: Multistage Fusion Algorithm for Remotely Sensed Underwater Super-Resolution Image Enhancement and Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 16, 3640–3653. [Google Scholar] [CrossRef]
- Liu, H.; Zhang, C.; Fan, B.; Xu, J. Pro2Diff: Proposal Propagation for Multi-Object Tracking via the Diffusion Model. IEEE Trans. Image Process. 2024, 33, 6508–6520. [Google Scholar] [CrossRef] [PubMed]
- Xuan, S.; Li, S.; Han, M.; Wan, X.; Xia, G.-S. Object tracking in satellite videos by improved correlation filters with motion estimation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1074–1086. [Google Scholar] [CrossRef]
- Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar]
- Li, H.; Man, Y. Moving ship detection based on visual saliency for video satellite. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 1248–1250. [Google Scholar] [CrossRef]
- Wu, J.; Zhang, G.; Wang, T.; Jiang, Y. Satellite Video Point-target Tracking in Combination with Motion Smoothness Constraint and Grayscale Feature. J. Geomat. 2017, 46, 1135–1146. [Google Scholar] [CrossRef]
- Zhang, C.; Wang, C.; Song, J.; Xu, Y. Based Satellite Video Object Tracking: A Review. Remote Sens. 2022, 14, 3674. [Google Scholar] [CrossRef]
- Chen, Y.; Tang, Y.; Xiao, Y.; Yuan, Q.; Zhang, Y.; Liu, F.; He, J.; Zhang, L. Satellite Video Single Object Tracking: A Systematic Review and An Oriented Object Tracking Benchmark. J. Photogramm. Remote Sens. 2024, 210, 212–240. [Google Scholar] [CrossRef]
- Du, B.; Sun, Y.; Cai, S.; Wu, C.; Du, Q. Object tracking in satellite videos by fusing the kernel correlation filter and the three-framedifference algorithm. IEEE Geosci. Remote Sens. Lett. 2018, 15, 168–172. [Google Scholar] [CrossRef]
- Possegger, H.; Mauthner, T.; Bischof, H. In defense of color-based model-free tracking. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2113–2120. [Google Scholar]
- Zhou, H.; Yuan, Y.; Shi, C. Object tracking using sift features and mean shift. Comput. Vis. Image Underst. 2009, 113, 345–352. [Google Scholar] [CrossRef]
- Nummiaro, K.; Koller-Meier, E.; Van Gool, L. Object tracking with an adaptive color-based particle filter. In Pattern Recognition; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2002; Volume 2449, pp. 353–360. [Google Scholar]
- Grabner, H.; Bischof, H. On-line boosting and vision. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; pp. 260–267. [Google Scholar]
- Grabner, H.; Leistner, C.; Bischof, H. Semi-supervised on-line boosting for robust tracking. In Computer Vision—ECCV 2008; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; pp. 234–247. [Google Scholar]
- Hare, S.; Golodetz, S.; Saffari, A.; Vineet, V.; Cheng, M.M.; Hicks, S.L.; Torr, P.H. Struck: Structured output tracking with kernels. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 263–270. [Google Scholar] [CrossRef] [PubMed]
- Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 4293–4302. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25; MIT Press: Cambridge, MA, USA, 2012; p. 2. [Google Scholar]
- Held, D.; Thrun, S.; Savarese, S. Learning to track at 100 fps with deep regression networks. In Proceedings of the Computer Vision ECCV, Amsterdam, The Netherlands, 8–16 October 2016; pp. 749–765. [Google Scholar]
- Li, M.; Zhang, H.; Wang, J.; Liu, Y. Research on ECO-HC Target Tracking Algorithm Based on Adaptive Template Update and Multi-Feature Fusion. IEEE Trans. Image Process. 2024, 23, 10758587. [Google Scholar] [CrossRef]
- Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar] [CrossRef]
- Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the circulant structure of tracking-by-detection with kernels. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 702–715. [Google Scholar]
- Kwon, J.; Lee, K. Visual Tracking Decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 1269–1276. [Google Scholar] [CrossRef]
- Deng, R.; Cui, C.; Remedios, L.W.; Bao, S.; Womick, R.M.; Chiron, S.; Li, J.; Roland, J.T.; Lau, K.S.; Liu, Q.; et al. Cross-scale Multi-instance Learning for Pathological Image Diagnosis. Med. Image Anal. 2024, 82, 103124. [Google Scholar] [CrossRef] [PubMed]
- Voigtlaender, P.; Luiten, J.; Torr, P.H.; Leibe, B. Siam R-CNN: Visual Tracking by Re-Detection. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6577–6587. [Google Scholar] [CrossRef]
Reference | Methods | Research Gap |
---|---|---|
[7] | HOG | HOG features describe the object by calculating the gradient direction and size, and the gradient direction is based on a fixed coordinate system. Cannot automatically adapt to such a rotation. |
[6] | KCF | The KCF model based on HOG does not have the ability of rotation adaptation and does not have the mechanism to deal with object occlusion. |
[21] | ECO | ECO is better than KCF in some aspects, but there are still shortcomings in rotating target tracking and occlusion processing, and the computational complexity is high. |
RACFME | KCF [6] | ECO [22] | Siam R-CNN [27] | MEDIANFLOW [25] | MIL [26] | BOOSTING [17] | |
---|---|---|---|---|---|---|---|
AUC | 78.6% | 57.4% | 59.1% | 72.7% | 10.2% | 52.2% | 49.6% |
Precision score | 96.7% | 73.4% | 77.8% | 90.5% | 8.2% | 42.1% | 47.6% |
Success score | 95.0% | 70.9% | 73.2% | 86.2% | 8.2% | 51.7% | 48.1% |
FPS | 116 | 130 | 60 | 46 | 90 | 7 | 62 |
RACFME | KCF [6] | ECO [22] | Siam R-CNN [27] | MEDIANFLOW [25] | MIL [26] | BOOSTING [17] | |
---|---|---|---|---|---|---|---|
AUC | 72.8% | 48.7% | 54.6% | 70.2% | 52.8% | 67.5% | 59.8% |
Precision score | 89.1% | 57.2% | 67.1% | 85.9% | 57.4% | 67.4% | 62.2% |
Success score | 86.7% | 50.1% | 61.3% | 83.6% | 50.7% | 82.6% | 74.5% |
FPS | 114 | 127 | 51 | 44 | 104 | 8 | 65 |
RACFME | KCF [6] | ECO [22] | Siam R-CNN [27] | MEDIANFLOW [25] | MIL [26] | BOOSTING [17] | |
---|---|---|---|---|---|---|---|
AUC | 76.5% | 37.4% | 42.8% | 52.7% | 53.0% | 24.0% | 44.9% |
Precision score | 91.3% | 47.2% | 54.8% | 64.1% | 55.7% | 41.1% | 47.7% |
Success score | 92.0% | 45.8% | 52.4% | 62.7% | 61.9% | 42.7% | 48.4% |
FPS | 121 | 131 | 60 | 46 | 97 | 7 | 64 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wu, X.; Zhang, H.; Mei, C.; Wu, J.; Ai, H. RACFME: Object Tracking in Satellite Videos by Rotation Adaptive Correlation Filters with Motion Estimations. Symmetry 2025, 17, 608. https://doi.org/10.3390/sym17040608
Wu X, Zhang H, Mei C, Wu J, Ai H. RACFME: Object Tracking in Satellite Videos by Rotation Adaptive Correlation Filters with Motion Estimations. Symmetry. 2025; 17(4):608. https://doi.org/10.3390/sym17040608
Chicago/Turabian StyleWu, Xiongzhi, Haifeng Zhang, Chao Mei, Jiaxin Wu, and Han Ai. 2025. "RACFME: Object Tracking in Satellite Videos by Rotation Adaptive Correlation Filters with Motion Estimations" Symmetry 17, no. 4: 608. https://doi.org/10.3390/sym17040608
APA StyleWu, X., Zhang, H., Mei, C., Wu, J., & Ai, H. (2025). RACFME: Object Tracking in Satellite Videos by Rotation Adaptive Correlation Filters with Motion Estimations. Symmetry, 17(4), 608. https://doi.org/10.3390/sym17040608