Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios
Abstract
:1. Introduction
- We propose a light segmentation method for water scenarios, which utilizes a two-stream learning strategy. Except for the traditional context stream, a spatial stream is expanded to learn spatial details in low-level layers with no extra computation cost in the inference time.
- We introduce edge-prior information to different layers in both streams, which leads to object-level semantic learning and memorizing, and expands the perspectives of pixel-level visual modeling.
- Evaluation on MODS benchmark and USV Inland dataset demonstrate that our approach achieves compelling performance. Notably, we obtain a significant improvement with a much lower number of parameters than the best frame-grained method.
2. Related Works
2.1. Water Scenario Segmentation
2.2. Edge Detection
2.3. Lightweight Networks
3. Method
3.1. Network Architecture
- Encoder. ELNet follows the traditional encoder–decoder structure to obtain the segmentation result. The encoder in the context stream is extremely essential for feature extraction and latent feature analysis. We choose MobileNetV2 [27] as the backbone, and most of the original design in the network will be retained as possible. Except for the last module for classification, MobileNetV2 has a total of 17 calculation modules. It contains multiple bottleneck blocks, which vary in five scales to extract features from the original image progressively.
- Decoder. To minimize the total number of parameters, we utilize the representative feature maps in three stages: the first stage, the third stage and the last stage. It is believed that the features in the first stage preserve the fundamental pixel-level information such as shapes of objects, the features in the last stage preserve the most abstract semantic information, and the features in the third stage are guided by the edge detail from the inputs and ground truth. Therefore, the features in the second and fourth stage are not as vital as the features in the first, third and last stages. For the feature maps at the first and third stage, the features after the last convolution layer is abandoned, instead by the features when the number of channels keeps at the maximum. This selection is conducive to fully preserving enriching information obtained by the encoder, and contributes to favorable information for segmentation. We also design an ablation study to validate the rationality of this design, and the experimental results are given at Section 4.7. The decoding and fusion strategy in the decoder can be formulated as:
- Auxiliary detail guidance. As shown in Figure 1, we use the low-level features (the 3rd stage) to produce detail guidance via the edge information and segmentation ground truth. Specifically, the guidance comes from two perspectives: consistent with the features of edge-prior information calculated from the Laplacian operator and with the ground truth of segmentation after decoding. Based on the above description, a Convolution Head is raised to regularize the third-stage feature maps. This module is carried out by
- Network details. Table 1 shows the detailed structure of the proposed network ELNet.
3.2. Detail Guidance in Spatial Path
- Guide with edge prior information. The detail feature prediction is modeled as a small knowledge distillation task and a binary segmentation task. We first generate the edge features encoded from the input edge image by a Laplacian operator and guide the partial learning of the third-scale coding features, which learns the same information from input image pairs. It can be illustrated as
- Guide with ground truth. Then, another Conv Head is utilized to generate segmentation prediction with the whole third-stage feature map and the detail ground-truth, which guides the feature map of the low-level layer to learn more spatial details. As shown in Figure 1, this guidance can be formulated as
3.3. Total Loss Function
4. Experiments
4.1. Dataset and Benchmark
- MaSTr1325 [42]: Marine Semantic Segmentation Training Dataset (MaSTr1325) is specially used to develop obstacle detection methods for small coastal USVs. The dataset contains 1325 reality-captured images, which include obstacles, water surface, sky and unknown targets, covering a series of real conditions encountered in coastal surveillance missions. It captures a variety of weather conditions, which range from foggy, partly cloudy with sunrise, overcast to sunny, and visually diverse obstacles, which are shown in Figure 2. The image size of MaSTr1325 is .
- MODS benchmark. [41]: The goal of MODS is to benchmark segmentation-based and detection-based obstacle detection methods for the maritime domain, specifically for use in unmanned surface vehicles (USVs). For segmentation-based detection, the segmentation method classifies each pixel in a given sensor image into one of three classes: sky, water or obstacle. Additionally, the MaSTr1325 dataset was created specifically for training.
- USVInland [43]: Different from the condition on the sea, the inland river environment, which is relatively narrow and complex, often brings additional challenges to the positioning and perception of the USV. Compared to the emerged public datasets in the field of road automatic driving, such as KITTI [44], Oxford RobotCar [45] and nuScenes [46], the USVInland dataset undoubtedly fills the gap and opens a new situation for inland river unmanned ships. A total of 27 pieces of original data are collected. There are relatively low resolution () and high resolution () images in the water segmentation sub-dataset. The fully original data will be directly used for validation.
4.2. Evaluation Metrics
4.3. Implementation Details
4.4. Comparison with Related Segmentation Methods
- First, a two-stream learning strategy (context stream learning with spatial stream learning) is applied. Additionally, the spatial stream works only in the training stage, which reduces the computation cost and thus affects the speed in the inference time.
- Second, the backbone of the proposed network refers to the designs of lightweight networks such as MobileNet series, Interception serires, etc., which have been approved to have a faster speed than traditional CNN networks.
- Third, we select an asymmetric decoder rather than a symmetric one with encoder after experiments, which also contributes to the speed in the inference time. The experimental detail is discussed in Section 4.7.
4.5. Performance on the MODS Benchmark
4.6. Performance on USVInland Dataset
4.7. Ablation Study on the Number of Upsampling Blocks
- Type 1: features in stage 1, 5, which preserve the features of the highest- and lowest-level.
- Type 2: features in stage 3, 5, which preserve the features of the finest-guidance and lowest-level.
- Type 3: features in stage 1, 3, 5, which preserve the features of the highest-level, lowest-level and the finest-guidance.
- Type 4: features in stage 1, 2, 4, 5, which preserve the features apart from that of the finest-guidance.
- Type 5: features in stage 1, 2, 3, 4, 5, which classically preserve the features of all levels.
4.8. Ablation Study on Edge-Aware Modules
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Liu, Z.; Zhang, Y.; Yu, X.; Yuan, C. Unmanned surface vehicles: An overview of developments and challenges. Annu. Rev. Control 2016, 41, 71–93. [Google Scholar] [CrossRef]
- Kristan, M.; Kenk, V.S.; Kovačič, S.; Perš, J. Fast image-based obstacle detection from unmanned surface vehicles. IEEE Trans. Cybern. 2015, 46, 641–654. [Google Scholar] [CrossRef] [PubMed]
- Luo, J.; Etz, S.P. A physical model-based approach to detecting sky in photographic images. IEEE Trans. Image Process. 2002, 11, 201–212. [Google Scholar] [PubMed]
- Lipschutz, I.; Gershikov, E.; Milgrom, B. New methods for horizon line detection in infrared and visible sea images. Int. J. Comput. Eng. Res. 2013, 3, 1197–1215. [Google Scholar]
- Yan, Y.; Shin, B.; Mou, X.; Mou, W.; Wang, H. Efficient horizon detection on complex sea for sea surveillance. Int. J. Electr. Electron. Data Commun. 2015, 3, 49–52. [Google Scholar]
- Liang, D.; Liang, Y. Horizon detection from electro-optical sensors under maritime environment. IEEE Trans. Instrum. Meas. 2019, 69, 45–53. [Google Scholar] [CrossRef]
- Taipalmaa, J.; Passalis, N.; Zhang, H.; Gabbouj, M.; Raitoharju, J. High-resolution water segmentation for autonomous unmanned surface vehicles: A novel dataset and evaluation. In Proceedings of the 2019 IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP), Pittsburgh, PA, USA, 13–16 October 2019; pp. 1–6. [Google Scholar]
- Bovcon, B.; Kristan, M. Obstacle detection for usvs by joint stereo-view semantic segmentation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 5807–5812. [Google Scholar]
- Chen, X.; Liu, Y.; Achuthan, K. WODIS: Water obstacle detection network based on image segmentation for autonomous surface vehicles in maritime environments. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar] [CrossRef]
- Zhou, R.; Gao, Y.; Wu, P.; Zhao, X.; Dou, W.; Sun, C.; Zhong, Y.; Wang, Y. Collision-Free Waterway Segmentation for Inland Unmanned Surface Vehicles. IEEE Trans. Instrum. Meas. 2022, 71, 1–16. [Google Scholar] [CrossRef]
- Vandaele, R.; Dance, S.L.; Ojha, V. Automated water segmentation and river level detection on camera images using transfer learning. In Proceedings of the DAGM German Conference on Pattern Recognition, Tübingen, Germany, 28 September–1 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 232–245. [Google Scholar]
- Kim, H.; Koo, J.; Kim, D.; Park, B.; Jo, Y.; Myung, H.; Lee, D. Vision-based real-time obstacle segmentation algorithm for autonomous surface vehicle. IEEE Access 2019, 7, 179420–179428. [Google Scholar] [CrossRef]
- Bovcon, B.; Kristan, M. A water-obstacle separation and refinement network for unmanned surface vehicles. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9470–9476. [Google Scholar]
- Yao, L.; Kanoulas, D.; Ji, Z.; Liu, Y. ShorelineNet: An efficient deep learning approach for shoreline semantic segmentation for unmanned surface vehicles. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 5403–5409. [Google Scholar]
- Lee, S.P.; Chen, S.C.; Peng, W.H. GSVNet: Guided spatially-varying convolution for fast semantic segmentation on video. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
- Lu, H.; Deng, Z. A Boundary-aware Distillation Network for Compressed Video Semantic Segmentation. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 5354–5359. [Google Scholar]
- Fan, M.; Lai, S.; Huang, J.; Wei, X.; Chai, Z.; Luo, J.; Wei, X. Rethinking BiSeNet for real-time semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9716–9725. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
- Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
- Bertasius, G.; Shi, J.; Torresani, L. Deepedge: A multi-scale bifurcated deep network for top-down contour detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4380–4389. [Google Scholar]
- Yu, Z.; Feng, C.; Liu, M.Y.; Ramalingam, S. Casenet: Deep category-aware semantic edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5964–5973. [Google Scholar]
- Shen, W.; Wang, X.; Wang, Y.; Bai, X.; Zhang, Z. Deepcontour: A deep convolutional feature learned by positive-sharing loss for contour detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3982–3991. [Google Scholar]
- Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. Mnasnet: Platform-aware neural architecture search for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; pp. 2820–2828. [Google Scholar]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
- Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
- Wu, B.; Dai, X.; Zhang, P.; Wang, Y.; Sun, F.; Wu, Y.; Tian, Y.; Vajda, P.; Jia, Y.; Keutzer, K. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; pp. 10734–10742. [Google Scholar]
- Wan, A.; Dai, X.; Zhang, P.; He, Z.; Tian, Y.; Xie, S.; Wu, B.; Yu, M.; Xu, T.; Chen, K.; et al. Fbnetv2: Differentiable neural architecture search for spatial and channel dimensions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19–20 June 2020; pp. 12965–12974. [Google Scholar]
- Dai, X.; Wan, A.; Zhang, P.; Wu, B.; He, Z.; Wei, Z.; Chen, K.; Tian, Y.; Yu, M.; Vajda, P.; et al. FBNetV3: Joint architecture-recipe search using predictor pretraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16276–16285. [Google Scholar]
- Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
- Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
- Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Bisenet: Bilateral segmentation network for real-time semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Bovcon, B.; Muhovič, J.; Vranac, D.; Mozetič, D.; Perš, J.; Kristan, M. MODS–A USV-oriented object detection and obstacle segmentation benchmark. IEEE Trans. Intell. Transp. Syst. 2021, 23, 13403–13418. [Google Scholar] [CrossRef]
- Bovcon, B.; Muhovič, J.; Perš, J.; Kristan, M. The mastr1325 dataset for training deep usv obstacle detection models. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 3431–3438. [Google Scholar]
- Cheng, Y.; Jiang, M.; Zhu, J.; Liu, Y. Are We Ready for Unmanned Surface Vehicles in Inland Waterways? The USVInland Multisensor Dataset and Benchmark. IEEE Robot. Autom. Lett. 2021, 6, 3964–3970. [Google Scholar] [CrossRef]
- Fritsch, J.; Kuehnl, T.; Geiger, A. A New Performance Measure and Evaluation Benchmark for Road Detection Algorithms. In Proceedings of the International Conference on Intelligent Transportation Systems (ITSC), Hague, The Netherlands, 6–9 October 2013. [Google Scholar]
- Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 Year, 1000 km: The Oxford RobotCar Dataset. Int. J. Robot. Res. IJRR 2017, 36, 3–15. [Google Scholar] [CrossRef]
- Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
- Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Operator | OS | T | N | S | IC | MC | OC | |
---|---|---|---|---|---|---|---|---|
Enc | conv2d | S/2 | - | 1 | 2 | 1 | - | 32 |
bottleneck | S/2 | 1 | 1 | 1 | 32 | - | 16 | |
bottleneck | S/4 | 6 | 2 | 2 | 16 | 96/144 | 24 | |
bottleneck | S/8 | 6 | 3 | 2 | 24 | 144/192 | 32 | |
bottleneck | S/8 | 6 | 4 | 2 | 32 | 192/384 | 64 | |
bottleneck | S/16 | 6 | 3 | 1 | 64 | 384/576 | 96 | |
bottleneck | S/32 | 6 | 3 | 2 | 96 | 576/960 | 160 | |
bottleneck | S/32 | 6 | 1 | 1 | 160 | 960 | 320 | |
Dec | up block | S/8 | - | 1 | 4 | 320 | - | 256 |
up block | S/2 | - | 1 | 4 | 448 | - | 64 | |
conv2d | S | - | 1 | 2 | 160 | - | 2 | |
Aux | conv block | S/8 | - | 1 | 1 | 3 | - | 16 |
conv block | S/8 | - | 1 | 1 | 192 | - | 4 |
Algorithms | Number of Parameters (M) | FPS on Gpu | FPS on Cpu |
---|---|---|---|
BiSeNet [39] | 13.42 | 36.12 | 4.87 |
WODIS [9] | 49.07 | 33.56 | 1.81 |
CollisionFree [10] | 100.36 | 9.89 | 0.21 |
Skip-ENet [12] | 0.75 | 59.82 | 11.24 |
WaSR [13] | 71.50 | 10.63 | 0.98 |
ShorelineNet [14] | 6.50 | 49.02 | 6.33 |
ELNet (Ours) | 4.86 | 45.21 | 6.95 |
Algorithms | TP | FP | FN | Pr [%] | Re [%] | F-Score [%] | |
---|---|---|---|---|---|---|---|
BiSeNet [39] | 12 (98.4) | 51,045 | 33,152 | 1443 | 60.6 | 97.3 | 74.7 |
WODIS [9] | 18 (97.1) | 49,966 | 87,651 | 2522 | 36.3 | 95.2 | 52.6 |
CollisionFree [10] | 53 (91.7) | 45,528 | 14,797 | 6960 | 75.5 | 86.7 | 80.7 |
Skip-ENet [12] | 25 (95.8) | 48,786 | 178,013 | 3702 | 21.5 | 92.9 | 34.9 |
WaSR [13] | 11 (98.6) | 51,607 | 85,374 | 881 | 37.7 | 98.3 | 54.5 |
ShorelineNet [14] | 12 (98.4) | 49,643 | 131,130 | 2845 | 27.5 | 94.6 | 42.6 |
ELNet (Ours) | 11 (98.5) | 51,429 | 29,318 | 1156 | 63.7 | 97.8 | 75.1 |
Algorithms | Pr [%] | Re [%] | F-Score |
---|---|---|---|
BiSeNet [39] | 97.89 | 87.77 | 93.04 |
WODIS [9] | 96.14 | 88.84 | 92.68 |
CollisionFree [10] | 93.64 | 75.96 | 84.88 |
Skip-ENet [12] | 96.81 | 76.78 | 86.65 |
WaSR [13] | 97.96 | 83.23 | 90.55 |
ShorelineNet [14] | 96.60 | 81.17 | 88.83 |
ELNet (Ours) | 97.82 | 88.69 | 93.96 |
Params | Pr [%] | Re [%] | F-Score [%] | ||
---|---|---|---|---|---|
1 | 8.34 M | 14 (98.0) | 60.6 | 95.9 | 64.7 |
2 | 4.42 M | 28 (95.3) | 20.9 | 92.3 | 34.3 |
3 | 4.87 M | 11 (98.5) | 63.7 | 97.8 | 75.1 |
4 | 6.05 M | 10 (98.6) | 63.4 | 97.7 | 75.2 |
5 | 5.17 M | 14 (98.0) | 60.2 | 96.6 | 69.5 |
Fusion | Auxiliary | F-Score | Re [%] | Pr [%] |
---|---|---|---|---|
× | × | 96.96 | 88.12 | 93.99 |
✓ | × | 97.32 | 88.37 | 93.97 |
× | ✓ | 97.36 | 88.42 | 93.88 |
✓ | ✓ | 97.82 | 88.69 | 93.96 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Han, W.; Zhao, B.; Luo, J. Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios. Sensors 2023, 23, 4789. https://doi.org/10.3390/s23104789
Han W, Zhao B, Luo J. Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios. Sensors. 2023; 23(10):4789. https://doi.org/10.3390/s23104789
Chicago/Turabian StyleHan, Wei, Binyu Zhao, and Jun Luo. 2023. "Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios" Sensors 23, no. 10: 4789. https://doi.org/10.3390/s23104789
APA StyleHan, W., Zhao, B., & Luo, J. (2023). Towards Smaller and Stronger: An Edge-Aware Lightweight Segmentation Approach for Unmanned Surface Vehicles in Water Scenarios. Sensors, 23(10), 4789. https://doi.org/10.3390/s23104789