Enhancing Situational Awareness of Helicopter Pilots in Unmanned Aerial Vehicle-Congested Environments Using an Airborne Visual Artificial Intelligence Approach
Abstract
:1. Introduction
- Collision Incidents: According to the Federal Aviation Administration (FAA), reports of UAVs interfering with manned aircraft systems have increased. In 2020, there were 1660 drone sightings reported to the FAA by pilots, citizens, and law enforcement, with numerous near-miss incidents that involved helicopters and small planes [9].
- Safety Concerns: The European Union Aviation Safety Agency (EASA) in its 2021 annual safety review [10] reported that there is a growing concern for potential collisions between unmanned aircraft systems and other aircraft in the area as a result of increasing accessibility to these systems.
- Technological Advancements: The proliferation of new and cheaper drone technologies can cause an overwhelmed sky, potentially violating safety regulations and increasing the cognitive load on pilots [11].
2. Related Work
2.1. Object Detection
2.2. Depth Estimation
2.3. Collision Prediction
2.4. Alert System
2.5. Contributions
3. Methodology
- Image/Video Acquisition Unit: The image and video acquisition unit is made of stereo cameras which are mounted on the helicopter, and these cameras are the ones that capture real-time video feeds of the surrounding airspace or environment. This subsystem collects image/video data and sends them to the data pre-processing unit. It consists of two stereo cameras, and they allow the execution of deep learning-based computer stereo vision, which is essential for depth estimation using the StereoNet model.
- Image/Video Pre-processing Unit: This unit takes the captured raw image/video data acquired by the stereo cameras and carries out its initial processing. This includes using pre-calibrated data to correct distortions and carrying out synchronization to make sure that stereo images are synchronized for accurate depth estimation.
- Image/Video Processing unit: The image and video processing unit is the core part of the entire system. This is because it contains the following modules, which are the basis of the entire system:
- Object detection module: this module uses the YOLO algorithm to detect UAV objects in the images.
- Depth Estimation module: this utilizes StereoNet to compute disparity maps, which are later converted to depth maps.
- Collision prediction module: this module implements Long Short-Term Memory (LSTM) to predict likely UAV collisions based on their respective trajectories.
- Threshold-based Alert Generation Module: this module initiates the alerts subsequent to the collision prediction made by the collision prediction module and when a distance threshold is reached.
- Monitor and Human–Machine Interface (HMI): this serves as a visual display for detected UAVs, their distances, and collision alerts. The HMI consists of the following:
- A visual display: this shows real-time video feeds indicating detected UAVs and their respective distances.
- Alert indicators: these are visual and audible indicators for collision alerts.
- Pilot interface: this allows pilots to interact with the overall system and also allows them to make any adjustments if required.
3.1. Helicopter Flying Environment Generation
3.2. Training and Testing Data
3.2.1. StereoNet Training Data
3.2.2. StereoNet Testing Data
3.2.3. LSTM Training Data
3.2.4. LSTM Testing Data
4. Results
4.1. Object Detection Results
4.2. Depth Estimation Results
4.3. Collision Prediction and Threshold-Based Alerts
- Drone Detected—Front Left! Collision in T seconds.
- Drone Detected—Front Right! Collision in T seconds.
4.4. Discussion
- The approach uses cameras as the main sensor for the detection of other drones. Cameras are sensitive to illumination conditions, which can result in the algorithm generating many false positives, thereby increasing the cognitive load on pilots. Here, other sensor alternatives or sensor fusion techniques are required to ensure robust and stable solutions.
- The prediction algorithm is based on linear mission profiles, such that the predicted trajectory and time to collision can be estimated accurately. However, missions with different maneuvers require different treatment with a richer dataset.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
Bi-LSTM | Bi-Directional Long Short-Term Memory |
CNN | Convolutional Neural Network |
EFA | European Aviation Safety Agency |
ESA | European Space Agency |
FAA | Federal Aviation Administration |
FPS | Frames Per Second |
GDPR | General Data Protection Regulation |
GPU | Graphics Processing Unit |
GRU | Gated Recurrent Unit |
HMI | Human Machine Interface |
LSTM | Long Short-Term Memory |
MAE | Mean Absolute Error |
RGB | Red Green Blue |
RNN | Recurrent Neural Network |
SIFT | Scale-Invariant Feature Transform |
SSD | Single Shot Multi-box Detector |
SVM | Support Vector Machine |
TTC | Time-To-Collision |
URDF | Unified Robot Description Format |
UAV | Unmanned Aerial Vehicle |
YOLO | You Only Look Once |
References
- Che Man, M.H.; Low, K.H. Damage severity prediction of helicopter windshields caused by a collision with a small unmanned aerial vehicle (sUAV). In Proceedings of the AIAA Aviation 2021 Forum, Virtual, 2–6 August 2021; p. 3001. [Google Scholar]
- Wallace, R.J.; Winter, S.R.; Rice, S.; Kovar, D.C.; Lee, S.A. Three case studies on small uncrewed aerial systems near midair collisions with aircraft: An evidence-based approach for using objective uncrewed aerial systems detection technology. SAE Int. J. Aerosp. 2023, 16, 263–278. [Google Scholar] [CrossRef]
- Paul, S.; Patterson, S.; Varela, C.A. Collaborative situational awareness for conflict-aware flight planning. In Proceedings of the 2020 AIAA/IEEE 39th Digital Avionics Systems Conference (DASC), Virtual, 11–15 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–10. [Google Scholar]
- Endsley, M.R. Supporting situation awareness in aviation systems. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation, Orlando, FL, USA, 12–15 October 1997; IEEE: Piscataway, NJ, USA, 1997; Volume 5, pp. 4177–4181. [Google Scholar]
- Adams, M.J.; Pew, R.W. Situational awareness in the commercial aircraft cockpit: A cognitive perspective. In Proceedings of the 9th IEEE/AIAA/NASA Conference on Digital Avionics Systems, Virginia Beach, VA, USA, 16–18 October 1990; IEEE: Piscataway, NJ, USA, 1990; pp. 519–524. [Google Scholar]
- Blasch, E.P. Learning attributes for situational awareness in the landing of an autonomous airplane. In Proceedings of the 16th DASC. AIAA/IEEE Digital Avionics Systems Conference. Reflections to the Future, Irvine, CA, USA, 26–30 October 1997; IEEE: Piscataway, NJ, USA, 1997; Volume 1. [Google Scholar]
- Andreeva-Mori, A.; Ohga, K.; Kobayashi, K.; Yoshida, A.; Takeichi, N. Manned aircraft situation awareness technology for low-altitude predicted trajectory conformance. In Proceedings of the 2022 IEEE/AIAA 41st Digital Avionics Systems Conference (DASC), Portsmouth, VA, USA, 18–22 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
- Jiang, L.; Jayatilaka, A.; Nasim, M.; Grobler, M.; Zahedi, M.; Babar, M.A. Systematic literature review on cyber situational awareness visualizations. IEEE Access 2022, 10, 57525–57554. [Google Scholar] [CrossRef]
- Howard, R.O. FAA unmanned aircraft systems (UAS) sighting reports: A preliminary survey. In Proceedings of the AIAA Aviation 2023 Forum, San Diego, CA, USA, 12–16 June 2023; p. 4099. [Google Scholar]
- Konert, A.; Kasprzyk, P. UAS Safety Operation—Legal issues on reporting UAS incidents. J. Intell. Robot. Syst. 2021, 103, 51. [Google Scholar] [CrossRef] [PubMed]
- Kasprzyk, P.J.; Konert, A. Reporting and investigation of Unmanned Aircraft Systems (UAS) accidents and serious incidents. Regulatory perspective. J. Intell. Robot. Syst. 2021, 103, 3. [Google Scholar] [CrossRef]
- Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Lu, Y.; Zhou, D.; Le, Q.V.; et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17182–17191. [Google Scholar]
- Katrakazas, C.; Quddus, M.; Chen, W.H.; Deka, L. Real-time motion planning methods for autonomous on-road driving: State-of-the-art and future research directions. Transp. Res. Part C Emerg. Technol. 2015, 60, 416–442. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886–893. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
- Han, X.; Chang, J.; Wang, K. You only look once: Unified, real-time object detection. Procedia Comput. Sci. 2021, 183, 61–72. [Google Scholar] [CrossRef]
- Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
- Kumar, A.; Perrusquía, A.; Al-Rubaye, S.; Guo, W. Wildfire and smoke early detection for drone applications: A light-weight deep learning approach. Eng. Appl. Artif. Intell. 2024, 136, 108977. [Google Scholar] [CrossRef]
- Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach; Pearson: London, UK, 2016. [Google Scholar]
- Valsaraj, A.; Barik, A.; Vishak, P.; Midhun, K. Stereo vision system implemented on FPGA. Procedia Technol. 2016, 24, 1105–1112. [Google Scholar] [CrossRef]
- Chalam, Y. Adaptive Control Systems: Techniques and Applications; Routledge: London, UK, 2017. [Google Scholar]
- Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-end learning of geometry and context for deep stereo regression. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 66–75. [Google Scholar]
- Khamis, S.; Fanello, S.; Rhemann, C.; Kowdle, A.; Valentin, J.; Izadi, S. Stereonet: Guided hierarchical refinement for real-time edge-aware depth prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 573–590. [Google Scholar]
- Luo, C.; McClean, S.I.; Parr, G.; Teacy, L.; De Nardi, R. UAV position estimation and collision avoidance using the extended Kalman filter. IEEE Trans. Veh. Technol. 2013, 62, 2749–2762. [Google Scholar] [CrossRef]
- Mohan, A.T.; Gaitonde, D.V. A deep learning based approach to reduced order modeling for turbulent flow control using LSTM neural networks. arXiv 2018, arXiv:1804.09269. [Google Scholar]
- Zhang, Y.; Jia, Z.; Dong, C.; Liu, Y.; Zhang, L.; Wu, Q. Recurrent lstm-based uav trajectory prediction with ads-b information. In Proceedings of the GLOBECOM 2022–2022 IEEE Global Communications Conference, Rio de Janeiro, Brazil, 4–8 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
- Yan, L.; Zhou, J.; Yang, K. Control-Aware Trajectory Predictions for Communication-Efficient Drone Swarm Coordination in Cluttered Environments. arXiv 2024, arXiv:2401.12852. [Google Scholar]
- Hu, P.; Tong, J.; Wang, J.; Yang, Y.; de Oliveira Turci, L. A hybrid model based on CNN and Bi-LSTM for urban water demand prediction. In Proceedings of the 2019 IEEE Congress on Evolutionary Computation (CEC), Wellington, New Zealand, 10–13 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1088–1094. [Google Scholar]
- Li, Y.; Chai, S.; Ma, Z.; Wang, G. A hybrid deep learning framework for long-term traffic flow prediction. IEEE Access 2021, 9, 11264–11271. [Google Scholar] [CrossRef]
- Perrusquía, A.; Guo, W. Reservoir computing for drone trajectory intent prediction: A physics informed approach. IEEE Trans. Cybern. 2024, 54, 4939–4948. [Google Scholar] [CrossRef] [PubMed]
- Ali, A.M.; Perrusquía, A.; Guo, W.; Tsourdos, A. Flight plan optimisation of unmanned aerial vehicles with minimised radar observability using action shaping proximal policy optimisation. Drones 2024, 8, 546. [Google Scholar] [CrossRef]
- Lin, T.T.; Weng, L.K.; Tsai, A.C. Object Tracking and Collision Avoidance Using Particle Filter and Vector Field Histogram Methods. In Proceedings of the 2014 ASABE – CSBE/SCGAB Annual International Meeting, Montreal, QC, Canada, 13–16 July 2014; American Society of Agricultural and Biological Engineers: St. Joseph, MI, USA, 2014; p. 1. [Google Scholar]
- Navarro González, J. People Exact-Tracking Using a Parrot AR. Drone 2.0. Bachelor’s Thesis, Universitat Politècnica de Catalunya, Barcelona, Spain, 2015. [Google Scholar]
- Rigoni, G.; Pinotti, C.M.; Das, D.; Das, S.K. Delivery with UAVs: A simulated dataset via ATS. In Proceedings of the 2022 IEEE 95th Vehicular Technology Conference: (VTC2022-Spring), Helsinki, Finland, 19–22 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
- Wang, X.; Lv, F.; Li, L.; Yi, Z.; Jiang, Q. A novel optimized tiny YOLOv3 algorithm for the identification of objects in the lawn environment. Sci. Rep. 2022, 12, 15124. [Google Scholar] [CrossRef] [PubMed]
- Oguine, K.J.; Oguine, O.C.; Bisallah, H.I. Yolo v3: Visual and real-time object detection model for smart surveillance systems (3s). In Proceedings of the 2022 5th Information Technology for Education and Development (ITED), Abuja, Nigeria, 1–3 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–8. [Google Scholar]
- Baade, A.; Peng, P.; Harwath, D. Mae-ast: Masked autoencoding audio spectrogram transformer. arXiv 2022, arXiv:2203.16691. [Google Scholar]
- Poggi, M.; Tosi, F.; Batsos, K.; Mordohai, P.; Mattoccia, S. On the synergies between machine learning and binocular stereo for depth estimation from images: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5314–5334. [Google Scholar] [CrossRef] [PubMed]
- Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th international conference on pattern recognition, Istanbul, Turkey, 23–26 August 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2366–2369. [Google Scholar]
- Fraser, B.; Perrusquía, A.; Panagiotakopoulos, D.; Guo, W. A deep mixture of experts network for drone trajectory intent classification and prediction using non-cooperative radar data. In Proceedings of the 2023 IEEE Symposium Series on Computational Intelligence (SSCI), Mexico City, Mexico, 5–8 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
- Perrusquía, A.; Guo, W.; Fraser, B.; Wei, Z. Uncovering drone intentions using control physics informed machine learning. Commun. Eng. 2024, 3, 36. [Google Scholar] [CrossRef]
- Mustafah, Y.M.; Azman, A.W.; Akbar, F. Indoor UAV positioning using stereo vision sensor. Procedia Eng. 2012, 41, 575–579. [Google Scholar] [CrossRef]
Model | MAE | RMSE |
---|---|---|
GRU | 0.0828 | 0.1292 |
LSTM | 0.0820 | 0.1284 |
Bi-LSTM | 0.0811 | 0.1281 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mugabe, J.; Wisniewski, M.; Perrusquía, A.; Guo, W. Enhancing Situational Awareness of Helicopter Pilots in Unmanned Aerial Vehicle-Congested Environments Using an Airborne Visual Artificial Intelligence Approach. Sensors 2024, 24, 7762. https://doi.org/10.3390/s24237762
Mugabe J, Wisniewski M, Perrusquía A, Guo W. Enhancing Situational Awareness of Helicopter Pilots in Unmanned Aerial Vehicle-Congested Environments Using an Airborne Visual Artificial Intelligence Approach. Sensors. 2024; 24(23):7762. https://doi.org/10.3390/s24237762
Chicago/Turabian StyleMugabe, John, Mariusz Wisniewski, Adolfo Perrusquía, and Weisi Guo. 2024. "Enhancing Situational Awareness of Helicopter Pilots in Unmanned Aerial Vehicle-Congested Environments Using an Airborne Visual Artificial Intelligence Approach" Sensors 24, no. 23: 7762. https://doi.org/10.3390/s24237762
APA StyleMugabe, J., Wisniewski, M., Perrusquía, A., & Guo, W. (2024). Enhancing Situational Awareness of Helicopter Pilots in Unmanned Aerial Vehicle-Congested Environments Using an Airborne Visual Artificial Intelligence Approach. Sensors, 24(23), 7762. https://doi.org/10.3390/s24237762