PISCFF-LNet: A Method for Autonomous Flight of UAVs Based on Lightweight Road Extraction
Abstract
:1. Introduction
1.1. Overview
- A lightweight network based on prior-information assistance and context feature fusion is proposed, PISCFF-LNet, significantly reducing model parameters and latency, making it suitable for deployment on edge devices;
- A feature-fusion module is designed to integrate shallow and deep encoder features, enhancing the network’s ability to handle multi-scale information;
- A vision-assisted UAV autonomous-flight-control method is proposed, RENA, optimizing road-edge extraction with the ray-based octagonal neighborhood algorithm to achieve basic terrain-following flight;
- A high-resolution UAV road-semantic-segmentation dataset is constructed, DRS Road, providing standardized data support for related research.
1.2. Autonomous Flight of UAVs
1.3. Road Extraction
2. Materials and Methods
2.1. Dataset
2.1.1. DeepGlobal Road
2.1.2. DRS Road
- (1)
- Multi-season and multi-geographical environments: The data span different seasons and a variety of geographical environments (such as desert, mountainous areas, and urban roads), ensuring the representativeness and wide applicability of the dataset.
- (2)
- High resolution and precise calibration: The use of high-resolution P1 optical cameras and LiDAR cameras ensures the accurate capture of image details, providing reliable data support for subsequent road extraction and analysis tasks.
2.2. Road-Extraction Network PISCFF-LNet
2.2.1. Prior-Information-Assisted Branch Based on Binarized Images
2.2.2. Lightweight Spatial Information Branch and Context Information Branch
2.2.3. Dual-Branch Information Fusion Method Based on Attention Refinement Module
2.3. Drone Interfacing
2.3.1. Ray-Based Eight-Neighborhood Algorithm
- Left boundary: Detect the 8 key neighboring pixels in a counterclockwise direction (0 → 1 → 2 → 3 → 4 → 5 → 6 → 7)
- Right boundary: Detect the 8 key neighboring pixels in a clockwise direction (0 → 7 → 6 → 5 → 4 → 3 → 2 → 1)
Algorithm 1 Ray-based edge navigation algorithm (RENA) |
Require: Sensor image Ensure: Target coordinate , fitting curves
|
2.3.2. Flight Control
3. Result
3.1. Relevant Metrics
3.1.1. Intersection over Union IoU
3.1.2. F1-Score
3.1.3. Frames per Second—FPS
3.2. Experimental Result
3.3. Ablation Studies
3.4. Experiments on UAV
3.4.1. Simulation Environment Experiments
3.4.2. Real-Environment Experiment
3.4.3. Interpretation of Results
4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Valavanis, K.P. Advances in Unmanned Aerial Vehicles: State of the Art and the Road to Autonomy; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
- Elsanhoury, M.; Koljonen, J.; Välisuo, P.; Elmusrati, M.; Kuusniemi, H. Survey on recent advances in integrated GNSSs towards seamless navigation using multi-sensor fusion technology. In Proceedings of the 34th International Technical Meeting of the Satellite Division of the Institute of Navigation (ION GNSS+ 2021), Online, 20–24 September 2021; pp. 2754–2765. [Google Scholar]
- Lu, Y.; Xue, Z.; Xia, G.S.; Zhang, L. A survey on vision-based UAV navigation. Geo-Spat. Inf. Sci. 2018, 21, 21–32. [Google Scholar]
- Achtelik, M.; Achtelik, M.; Weiss, S.; Siegwart, R. Onboard IMU and monocular vision based control for MAVs in unknown in-and outdoor environments. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3056–3063. [Google Scholar]
- Joachims, T. 11: Making Large-Scale Support Vector Machine Learning Practical. In Advances in Kernel Methods: Support Vector Learning; Technical Report; Burges, C.J.C., Schölkopf, B., Smola, A.J., Eds.; The MIT Press: Cambridge, MA, USA, 1998; ISBN 9780262283199. [Google Scholar] [CrossRef]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar]
- Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN model-based approach in classification. In Proceedings of the on the Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE: OTM Confederated International Conferences, CoopIS, DOA, and ODBASE 2003, Catania, Italy, 3–7 November 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 986–996. [Google Scholar]
- Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 630–645. [Google Scholar]
- Vaswani, A. Attention is all you need. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
- Yurtkulu, S.C.; Şahin, Y.H.; Unal, G. Semantic segmentation with extended DeepLabv3 architecture. In Proceedings of the 2019 IEEE 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; pp. 1–4. [Google Scholar]
- Valanarasu, J.M.J.; Patel, V.M. Unext: Mlp-based rapid medical image segmentation network. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 23–33. [Google Scholar]
- Yun, B.; Peng, K.; Chen, B.M. Enhancement of GPS signals for automatic control of a UAV helicopter system. In Proceedings of the 2007 IEEE International Conference on Control and Automation, Guangzhou, China, 30 May–1 June 2007; pp. 1185–1189. [Google Scholar]
- Nemra, A.; Aouf, N. Robust INS/GPS sensor fusion for UAV localization using SDRE nonlinear filtering. IEEE Sens. J. 2010, 10, 789–798. [Google Scholar]
- Yang, T.; Li, P.; Zhang, H.; Li, J.; Li, Z. Monocular vision SLAM-based UAV autonomous landing in emergencies and unknown environments. Electronics 2018, 7, 73. [Google Scholar] [CrossRef]
- Qiu, Z.; Lin, D.; Jin, R.; Lv, J.; Zheng, Z. A global aruco-based lidar navigation system for uav navigation in gnss-denied environments. Aerospace 2022, 9, 456. [Google Scholar]
- Bachrach, A.; He, R.; Roy, N. Autonomous flight in unknown indoor environments. Int. J. Micro Air Veh. 2009, 1, 217–228. [Google Scholar]
- Sakthivel, P.; Anbarasu, B. Integration of vision and LIDAR for navigation of micro aerial vehicle. In Proceedings of the 2020 IEEE Third International Conference on Multimedia Processing, Communication & Information Technology (MPCIT), Shivamogga, India, 11–12 December 2020; pp. 14–18. [Google Scholar]
- Gageik, N.; Strohmeier, M.; Montenegro, S. An autonomous UAV with an optical flow sensor for positioning and navigation. Int. J. Adv. Robot. Syst. 2013, 10, 341. [Google Scholar]
- Arshad, M.A.; Khan, S.H.; Qamar, S.; Khan, M.W.; Murtza, I.; Gwak, J.; Khan, A. Drone Navigation Using Region and Edge Exploitation-Based Deep CNN. IEEE Access 2022, 10, 95441–95450. [Google Scholar] [CrossRef]
- Zhao, Y.; Zhang, J.; Zhang, C. Deep-learning based autonomous-exploration for UAV navigation. Knowl.-Based Syst. 2024, 297, 111925. [Google Scholar]
- Ren, Y.; Yu, Y.; Guan, H. DA-CapsUNet: A dual-attention capsule U-Net for road extraction from remote sensing imagery. Remote Sens. 2020, 12, 2866. [Google Scholar] [CrossRef]
- Li, J.; Liu, Y.; Zhang, Y.; Zhang, Y. Cascaded attention DenseUNet (CADUNet) for road extraction from very-high-resolution images. ISPRS Int. J. Geo-Inf. 2021, 10, 329. [Google Scholar] [CrossRef]
- Zhou, L.; Zhang, C.; Wu, M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 182–186. [Google Scholar]
- Zhou, M.; Sui, H.; Cheng, X. A novel dominant feature driven urban road extraction method. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 1–4. [Google Scholar]
- Li, X.; Wang, Y.; Zhang, L.; Liu, S.; Mei, J.; Li, Y. Topology-enhanced urban road extraction via a geographic feature-enhanced network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8819–8830. [Google Scholar]
- Mei, J.; Li, R.J.; Gao, W.; Cheng, M.M. CoANet: Connectivity attention network for road extraction from satellite imagery. IEEE Trans. Image Process. 2021, 30, 8540–8552. [Google Scholar]
- Tan, H.; Xu, H.; Dai, J. BSIRNet: A road extraction network with bidirectional spatial information reasoning. J. Sens. 2022, 2022, 6391238. [Google Scholar]
- Chen, T.; Jiang, D.; Li, R. Swin transformers make strong contextual encoders for VHR image road extraction. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 3019–3022. [Google Scholar]
- Zhang, Z.; Miao, C.; Liu, C.; Tian, Q. DCS-TransUperNet: Road segmentation network based on CSwin transformer with dual resolution. Appl. Sci. 2022, 12, 3511. [Google Scholar] [CrossRef]
- Liu, B.; Ding, J.; Zou, J.; Wang, J.; Huang, S. LDANet: A lightweight dynamic addition network for rural road extraction from remote sensing images. Remote Sens. 2023, 15, 1829. [Google Scholar] [CrossRef]
- Wang, R.; Cai, M.; Xia, Z. A lightweight high-resolution RS image road extraction method combining multi-scale and attention mechanism. IEEE Access 2023, 11, 108956–108966. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar]
- Xie, G.; He, L.; Lin, Z.; Zhang, W.; Chen, Y. Lightweight optical remote sensing image road extraction based on L-DeepLabv3+. Laser J. 2024, 45, 111–117. [Google Scholar]
- Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
- Qu, G.; Wu, Y.; Lv, Z.; Zhao, D.; Lu, Y.; Zhou, K.; Tang, J.; Zhang, Q.; Zhang, A. Road-MobileSeg: Lightweight and Accurate Road extraction model from Remote sensing images for Mobile devices. Sensors 2024, 24, 531. [Google Scholar] [CrossRef] [PubMed]
- Lin, K.; Zhang, S.; Qin, Z. ConvPose: An efficient human pose estimation method based on ConvNeXt. In Proceedings of the 5th International Conference on Computer Science and Software Engineering, Guilin, China, 21–23 October 2022; pp. 80–84. [Google Scholar]
- Yu, C.; Gao, C.; Wang, J.; Yu, G.; Shen, C.; BiSeNet, N.S. Bisenet v2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vis. 2021, 129, 3051–3068. [Google Scholar] [CrossRef]
- Fan, M.; Lai, S.; Huang, J.; Wei, X.; Chai, Z.; Luo, J.; Wei, X. Rethinking bisenet for real-time semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9716–9725. [Google Scholar]
- Amovlab. Prometheus Autonomous UAV Opensource Project. Available online: https://github.com/amov-lab/Prometheus (accessed on 1 September 2024).
Algorithm | Time Complexity | Space Complexity | Processing Time (ms) |
---|---|---|---|
Eight-Neighbor Algorithm | 9200 | ||
RENA | O(1) | 10 |
Methods | DeepGlobe Road | DRS Road | Model Index | ||||
---|---|---|---|---|---|---|---|
IoU | F1-Score | IoU | F1-Score | Params (M) | FLOPs (G) | FPS | |
DeepLabV3plus_MobileNetV2 | 65.71% | 79.31% | 88.58% | 93.95% | 5.81 | 26.42 | 2.22 |
DeepLabV3plus_xception | 64.36% | 78.31% | 87.31% | 93.23% | 54.71 | 83.42 | 0.73 |
UNeXt [14] | 65.70% | 79.30% | 88.54% | 93.92% | 1.47 | 2.29 | 7.21 |
BiSeNetV2 [41] | 64.98% | 78.77% | 88.01% | 93.62% | 3.61 | 12.91 | 3.75 |
STDC1 [42] | 49.38% | 66.11% | 83.86% | 91.22% | 14.23 | 23.52 | 4.72 |
PISCFF-LNet [ours] | 66.86% | 80.14% | 89.61% | 94.52% | 2.31 | 5.38 | 5.39 |
No. | PIA | CIB-SIB | DI-ARM | Seg | Accuracy | IoU | Precision | Recall | F1-Score |
---|---|---|---|---|---|---|---|---|---|
1 | 98.21% | 64.87% | 78.77% | 78.61% | 78.69% | ||||
2 | Y | 98.23% | 65.05% | 79.30% | 78.35% | 78.82% | |||
3 | Y | Y | 98.20% | 64.54% | 79.02% | 77.89% | 78.45% | ||
4 | Y | Y | Y | 98.32% | 66.69% | 80.20% | 79.83% | 80.02% | |
5 | Y | Y | Y | Y | 98.34% | 66.86% | 80.81% | 79.48% | 80.14% |
Name | Configuration |
---|---|
CPU | Intel(R) Core(TM) i7-12650H 2.30 GHz (Intel Corporation, Santa Clara, CA, USA) |
GPU | NVIDIA RTX 4060 (NVIDIA Corporation, Santa Clara, CA, USA) |
Memory | 32 GB |
Operating System | Ubuntu 20.04.6 LTS |
Data Processing | OpenCV, PIL |
Python Version | Python 3.9.13 |
Deep Learning Framework | PyTorch 2.1.0 |
CUDA Version | CUDA 11.8 |
Name | Specification/Model |
---|---|
Frame | MFP_V1 410 mm |
Onboard Computer | Viobot RK3588 |
Battery | FB45 4S 5000 mAh |
Remote Controller | AMOVLAB QE-2 |
Flight Controller | Pixhawk 6C |
Video Transmitter | Mini Homer |
Motor | 2312 960 kv |
Propeller | 0-inch |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhu, Y.; Zhang, T.; Wu, A.; Shi, G. PISCFF-LNet: A Method for Autonomous Flight of UAVs Based on Lightweight Road Extraction. Drones 2025, 9, 226. https://doi.org/10.3390/drones9030226
Zhu Y, Zhang T, Wu A, Shi G. PISCFF-LNet: A Method for Autonomous Flight of UAVs Based on Lightweight Road Extraction. Drones. 2025; 9(3):226. https://doi.org/10.3390/drones9030226
Chicago/Turabian StyleZhu, Yuanxu, Tianze Zhang, Aiying Wu, and Gang Shi. 2025. "PISCFF-LNet: A Method for Autonomous Flight of UAVs Based on Lightweight Road Extraction" Drones 9, no. 3: 226. https://doi.org/10.3390/drones9030226
APA StyleZhu, Y., Zhang, T., Wu, A., & Shi, G. (2025). PISCFF-LNet: A Method for Autonomous Flight of UAVs Based on Lightweight Road Extraction. Drones, 9(3), 226. https://doi.org/10.3390/drones9030226