Next Article in Journal
Dynamic Visual Measurement of Driver Eye Movements
Next Article in Special Issue
Preparation and Characterization of Developed CuxSn1−xO2 Nanocomposite and Its Promising Methane Gas Sensing Properties
Previous Article in Journal
Chemical Sensors Generated on Wafer-Scale Epitaxial Graphene for Application to Front-Line Drug Detection
Previous Article in Special Issue
Gas Sensors Based on Mechanically Exfoliated MoS2 Nanosheets for Room-Temperature NO2 Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autonomous Visual Perception for Unmanned Surface Vehicle Navigation in an Unknown Environment

1
School of Navigation, Wuhan University of Technology, Wuhan 430063, China
2
Department of Electrical and Computer Engineering, George Mason University, Fairfax, VA 22030, USA
3
Hubei Key Laboratory of Inland Shipping Technology, Wuhan University of Technology, Wuhan 430063, China
4
National Engineering Research Center for Water Transport Safety, Wuhan University of Technology, Wuhan 430063, China
5
Institute of Ocean Information Technology, Shandong Jiaotong University, Weihai 250357, China
6
School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
7
School of Transportation, Wuhan University of Technology, Wuhan 430063, China
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(10), 2216; https://doi.org/10.3390/s19102216
Submission received: 9 April 2019 / Revised: 4 May 2019 / Accepted: 7 May 2019 / Published: 14 May 2019
(This article belongs to the Special Issue Gas Sensors and Smart Sensing Systems)

Abstract

:
Robust detection and recognition of water surfaces are critical for autonomous navigation of unmanned surface vehicles (USVs), since any none-water region is likely an obstacle posing a potential danger to the sailing vehicle. A novel water region visual detection method is proposed in this paper. First, the input image pixels are clustered into different regions and each pixel is assigned a label tag and a confidence value by adaptive multistage segmentation algorithm. Then the resulting label map and associated confidence map are fed into a convolutional neural network (CNN) as training samples to train the network online. Finally, the online trained CNN is used to segment the input image again but with greater precision and stronger robustness. Compared with other deep-learning image segmentation algorithms, the proposed method has two advantages. Firstly, it dispenses with the need of manual labeling training samples which is a costly and painful task. Secondly, it allows real-time online training for CNN, making the network adaptive to the navigational environment. Another contribution of this work relates to the training process of neuro network. An effective network training method is designed to learn from the imperfect training data. We present the experiments in the lake with a various scene and demonstrate that our proposed method could be applied to recognize the water region in the unknown navigation environment automatically.

1. Introduction

The past decade of research in marine and field robotics gave rise to the development of a new class of small-sized USVs. Currently, the USV is playing more and more important roles in maritime applications, such as meteorological monitoring, maritime search and rescue, geophysical exploration, environmental monitoring [1,2]. As a vision sensor, the video cameras are equipped in almost all of USVs to accomplish a variety of missions [3,4,5,6]. Although by itself camera is data-rich, how to extract useful information from millions or billions of raw pixels remains to be a challenging problem. One of those specific to USVs is water region segmentation, i.e., assigning each pixels a meaningful label such as water, sky, shore, or other vehicles in view, even a coarse water-none-water labeling is very helpful to the auto-navigation task of USV.
The water region recognition in the image is of great importance in many applications, such as obstacle detection and avoidance, path planning [7]. The target search could be constrained in the specific region by the water region segmentation to reduce the computation. In the past, a lot of researchers directly focus on the detection of the water-shore or water-sky boundary by analyzing low level image features such as color, edge or texture [8,9,10,11]. In [11], the Otsu’s thresholding algorithm was applied to threshold the images into binary images. Then, the sea-sky line was extracted with Hough transform as the longest straight line. Zafarifar et al. [12] used a method based on edge and color to detect the horizon line in an image. In [13], a hierarchical horizon detection algorithm was proposed by using Canny edge detector and Hough transform to find major edges in the image, and the precise curvature of the horizon was captured by fine-level adjustment. Lu et al [14] designed a modified edge detection algorithm according to the traditional Canny algorithm. Least square optimization is also a staple regression method for line detection but this method is more sensitive to noise than the Hough transform based methods [15]. Wang et al. [16] sampled maximum or minimum gradient value points along each vertical line in the gradient image, and then used RANSAC to fit the horizon line.
With the invention of novel algorithms and great improvement of computing power, deep learning based algorithms has achieved remarkable performance in many computer vision tasks, such as visual recognition [17], video processing [18,19] and semantic segmentation [20]. It has already been applied to the USV in the navigation, obstacle avoidance and environment perception [21,22]. Semantic segmentation as one of the widest fields of the application improves the scene recognition capability of the USV. Instead of only focusing on the contour boundary, these methods classify all pixels of the image. Fefilatyev et al. [23] proposed to segment the sky and then detect the horizon line in the binary mask of the network output. Cheng proposed the SeNet [24] to segment sea-land region, which combined the segmentation task and edge detection task into an end-to-end deconvNet in a multi-task way. However, the methods based on the state-of-the-art deep learning heavily rely on the manual label training data set.
In this paper, we propose a novel online learning approach to recognize the water region for the USV in the unknown navigation environment. Our method is divided into two main steps. Firstly, the water surface region is predicted by the proposed multi-stage adaptive segmentation method. Secondly, the predicted map is used as labeled training samples to train our segmentation network. Then the online trained network is used to segment the input image with a post-processing of conditional random field (CRF) to refine and improve the result. The rest of the paper is organized as follows: Section 2 describes the proposed algorithm in detail; Experimental dataset and results are demonstrated and compared in Section 3; Section 4 concludes this work.

2. Proposed Method

In this section, the proposed method for the dynamic water surface segmentation is explained in detail. The diagram of the proposed algorithm is shown in Figure 1. The free safe water region is detected by an obstacle avoidance sensor. In our experiment, a lidar sensor is used to perceive the obstacle in front of the USV. Based on the spatial relationship between the lidar sensor and camera, we can determine whether there are obstacles in the special region of the image corresponding to the sensing range of the lidar. Then, each pixel in the special region of the image is confirmed whether they are obstacles or just water surfaces.
Meanwhile, the adaptive segmentation method is used to segment the image into different regions based on the similarity in color and texture. Different segmentation results are obtained by adjusting the sensitivity of two region merging. By adjusting the sensitivity of the region merging, we get multiple segmentation of the image. Then, the self-generated label and the corresponding weight map are generated according to the multi-stage segmentation and the free safe water region. Finally, the binary cross entropy loss function, which takes into consideration the diverse uncertainty in different regions, is designed to train the network. When the segmentation network is used for prediction, the conditional random field (CRF) is used to refine the network output.

2.1. Adaptive Segmentation

In this article, a modified and enhanced version of graph-based image segmentation technique adapted from algorithm proposed by Felzenszwalb [25] is used to segment the input image initially. Compared with the k-mean based segmentation method that takes the global color statistical characteristic into consideration, the graph-based method is based on the local color and texture feature. In a graph-based segmentation approach, each pixel is regarded as a separate region initially, then each of those small regions are merged into super region recursively until specified ending criteria is satisfied. The following is a detailed introduction of the region merging process.
G = ( V ,   E ) is an undirected graph with vertices v i V , the set of image pixels to be segmented, and edges e ( v i ,   v i ) E corresponding to pairs of neighboring pixels. Each edge e ( v i ,   v i ) has a corresponding weight w ( v i ,   v j ) , which is a non-negative measure of the dissimilarity between the two pixels connected by that edge (e.g., the difference in color, location or some other local attribute). A segmentation S is a partition of V into components such that each region C S corresponds to a connected region in a graph G = ( V , E ) , where E E . Ideally, edges between two pixels in the same region should have relatively low weight while edges between pixels in different region should have higher weight.
A predicate, D, is defined for measuring the evidence for a boundary between two regions in a segmentation. The predicate is based on comparing the dissimilarity between regions along the boundary of the two regions relative to the dissimilarity among neighboring regions within each of the two regions. The dissimilarity within each region C V is defined to be the largest edge weight of the minimum spanning tree in the region, MST ( C ,   E ) :
Int ( C ) = max e M S T ( C , E ) w ( e )
The external dissimilarity is defined by the difference between the two regions C 1 ,   C 2 V to be the minimum edge weight connecting the two regions:
Dif ( C 1 ,   C 2 ) = min v i C 1 , v j C 2 , ( v i , v j ) E w ( ( v i ,   v j ) )
Whether there is a boundary between a pair region is determined by checking if the difference between the regions, Dif ( C 1 ,   C 2 ) , is larger compared with the internal difference within at least one of the regions, Int ( C 1 ) and Int ( C 2 ) . The pairwise comparison is defined as:
D ( C 1 , C 2 ) = { t r u e i f   Dif ( C 1 ,   C 2 ) > MInt ( C 1 , C 2 ) f a l s e o t h e r w i s e
where Mint ( · ) is defined as:
MInt ( C 1 , C 2 ) = min ( I n t ( C 1 ) + τ ( C 1 ) , I n t ( C 2 ) + τ ( C 2 )
The function τ controls the degree to ensure that the difference between the two regions must be greater than their internal differences. To avoid the influence of the small region, the function τ is defined with a factor of the region size:
τ ( C ) = k | C |
where | C | denotes the size of C , and the parameter k is a constant.
The size of the segmented region is increased as k grows. Due to the various light intensity and visibility of the water environment, the algorithm has a poor performance with the constant parameter k. To make the algorithm robust to the different condition, the τ ( C ) is modified as:
τ ( C ) = k 0 · σ I | C |
where σ I is the mean standard deviation of the image pixel value in each channel:
σ I = 1 3 i = r , g , b σ i
σ i = ( p i P ¯ i ) 2 H · W
where H and W are the height and width of the image. By relating the controlling parameter k to the standard deviation of the image pixels, the modified algorithm is able to achieve better segmentation performance in adversary situation such as scene with low contrast or low visibility

2.2. Water Region Prediction Based on Multi Stage Segmentation

In order to classify the segmented region into water or none-water surface, a lidar device is used to sense a small region within close range of USV; the corresponding patch in the image is assumed to be water surface when there is no reflection signal from lidar. The segmented region of similar color and texture as this special region are then merged together and labeled as water surface; all the other segmented regions are labeled as none-water region:
c ( x , y ) = { 1 i f   p ( x m , y m ) ( s i s R ) ,   c ( x m , y m ) = 1 0 otherwise
The equation above describes the process of the water region prediction. s i is a region of the image segmentation S and p ( x ,   y ) s i . s R is the region that is corresponding to the sensing range of the lidar. If there is a point belonging to s R and confirmed as the water region, then any point in the region is regarded as the water surface. Otherwise, it’s regarded as the non-water-surface.
During the adaptive segmentation, the tendency of two regions merging into one region decreases as the value of the function τ increases. When τ is small, the segmentation is of fine grain and segmented regions are of small size. The prediction whether a region connected to the convince water surface is also the water region have high confidence. And there is a high possibility that many of the regions regarded as the non-water-region are indeed the water, i.e., a miss. On the other hand, when τ is large, the segmentation is coarse and the segmented regions are of large size. There is a high possibility that parts of the large regions connected to the convince water surface are non-water, i.e., a false alarm. To reduce the dependency on empirically adjusting parameters, multiple value of τ are used and combined to generate a probability map:
τ n ( C ) = f n k 0 · σ I | C |
f n = 1.5 n 1
where f n is a factor that controls the sensibility of the segmentation and n { 0 ,   1 , ,   9 } . Through the multistage segmentation with the different value of f n , each pixel of the image could have a set of labels as define by Equation (9), c ( x , y ) = { c 0 ,   c 1 , ,   c 10 } . Then the probability of the pixels that belongs to the water surface is defined as:
q ( x , y ) = 1 10 i = 0 10 c i

2.3. Segmentation Network

The label maps generated previously by multistage segmentation are used as training samples to train the segmentation network that is based on U-Net [26] which is a very popular end-to-end encoder-decoder network for semantic segmentation. The U-Net structure uses multiple feature maps in different layers and has a good performance in different applications.
Each layer in the encoder network performs convolution with a filter bank to produce a set of feature maps. The max-pooling with a 2 × 2 window and stride 2 is performed to achieve translation invariance over small spatial shifts in the input image. The output is sub-sampled by a factor of 2 in each max-pooling operation. This results in a large input image context for each pixel in the feature map. In order to use the available pre-trained model, the last fully connected layers of VGG16 [27] are eliminated and the fully convolutional layers of VGG16 are reserved as the encoder. Based on the feature map size, the convolution layers are divided into six stages. The detailed structure of the designed network is shown in Figure 2.

2.4. Online Training

A cross entropy loss function is used to train the network. N is defined as the number of image pixels used to update the network parameters. y ( x , y ) and y ^ ( x , y ) are defined as the labels of the pixel p ( x , y ) predicted by the proposed multi-stage adaptive segmentation method and the corresponding predicted probability from the network respectively. The binary cross entropy loss is then expressed as:
L = 1 N N [ y ( x , y ) l o g y ^ ( x , y ) + ( 1 y ( x , y ) ) l o g ( 1 y ^ ( x , y ) ) ]
As the label is auto-generated with uncertainty, the training with any wrongly labeled data results in network degradation. To improve the training efficiency, we emphasize on the pixels with high confidence during the training.
The value calculated by Equation (12) is interpreted as the possibility of being classified as water region, where value near 1 means confirmation of water region and 0 means confirmation of none-water region and any value near 0.5 rendering the prediction is unreliable. Based on the pixel prediction, a confidence map is generated to represent the reliability of each pixel prediction. The confidence map is estimated as:
w ( x , y ) = c o s ( 2 π · q ( x , y ) ) + 1 2
A cosine function is the mapping function. As q ( x , y ) varies from 0 to 1, w ( x , y ) has a high value at both ends and a low value in the middle. Thus, the confidence map is regarded as the weight map to address the problem of the reliability imbalance in different pixels. The binary cross entropy loss is modified as:
L = 1 N N w ( x , y ) [ y ( x , y ) l o g y ^ ( x , y ) + ( 1 y ( x , y ) ) l o g ( 1 y ^ ( x , y ) ) ]
The advantage of this loss function is that it puts more weight on pixels with greater label confidence and less weight on those unreliable pixels, hence reduces the degradation of network caused by bad training samples.

2.5. Segmentation Improvement

In the field of image multi-class object segmentation, the conditional random field (CRF) is an effective way to classify the pixel category by considering the relationship between a single pixel and all other pixels in an image [28]. In this paper, a method adopted from the work of Deeplab [29] that combined CNN and CRF, is used to further improve the output possibility map of the segmentation network. The energy function of the CRF model is:
E ( x ) = i φ i ( x i | y ) + i , j ϕ i , j ( x i , x j | y )
As in the standard CRF-based image semantic segmentation [30], each pixel is represented as a random variable. Each of these random variables takes a label from the set L = { l 1 , l 2 } , which represent the water region and the others. Let X = { X 1 , X 2 , , X N } denote the set of random variables corresponding to the image pixels. N is equal to H × W ( H is the Height and W is the width of the image). A labeling x refers to any possible assignment of labels to the random variables and takes values from the set L. In the given image, y represents the image pixel value (or the features map extracted from the small area around pixel i), and x i represents the random label of pixel i.

2.5.1. Unary Potentials

φ i ( x i | y ) = l o g P ( x i | y )
The unary potential is defined as the negative log of the likelihood of a label being assigned to pixel i. It can be treated as local classifiers and well defined by the output of the network model, a probability map for each class in each pixel.

2.5.2. Pairwise Potentials

The pairwise potential models the relationship among neighboring pixels and weighted by color similarity. The pairwise potential is defined as the following:
ϕ i , j ( x i , x j | y ) = u ( x i , x j ) [ ω 1 e x p ( p i p j 2 2 σ α 2 y i y j 2 2 σ β 2 ) + ω 2 e x p ( p i p j 2 2 σ γ 2 ) ]
where p i and p j are the pixel position. u ( x i , y j ) represents the compatibility of the labels assigned to variables x i and x j , where u ( x i , x j ) = 0 if x i = x j and 1 otherwise. The first term depends on both pixel positions and pixel interactions with the neighboring information. And the second term only depends on pixel positions.
Each random value of the CRF is initialized as the corresponding output from the CNN segmentation network described in Section 2.3. Then the optimization result is obtained by minimizing the above CRF energy E(x). CRF refines the local segmentation and improve the algorithm robustness by considering the color and spatial information.

3. Experimental Results

To verify the performance of the proposed method, several experiments and comparison were carried out based on the data captured by an experimental USV navigating in East Lake located in Wuhan, China. Figure 3 shows the experimental USV and the navigation scene. Table 1 shows the detail of the experimental USV and computer.

3.1. Adaptive Segmentation

The proposed adaptive segmentation is based on the graph-based segmentation algorithm. To demonstrate the robustness of our multistage adaptive segmentation algorithm, we compare our method with the other similar graph-based segmentation algorithm [25] which has a standard implementation in the popular OpenCV library. As shown in Formula (5), there is a parameter k controlling the grain level of the segmentation, so we run the algorithm multiple times with different k value and compared the results with that of the proposed method.
Four typical lake scenes were used as the testing case. For each scene, the minimum region size of segmentation was set as 100, but three different k value were tried, k = 600, k = 1200 and k = 2000. The result of different algorithms or parameters are tiled in Figure 4. As shown in the first and second rows of the figure, the segmentation is too coarse with a large k for low contrast scene. As shown in the third and fourth rows of the figure, the segmentation is too fine with a small k for normal contrast scene. However, our proposed method could segment all of these images properly without tuning any parameter, neither too fine nor too coarse in any case.

3.2. Process of the Training Data Generating

This experiment shows the process of the water region prediction. Based on the proposed adaptive segmentation algorithm, ten segmentations from fine to coarse are obtained. Based on the lidar information and the multistage segmentation result, the water region prediction map is generated by fusing maps from different stage as Equation (12), then the weight map is generated as Equation (14). Figure 5 shows the final prediction map and the corresponding weight map. The prediction maps and the weight maps are shown by the second and third row of the images, respectively.

3.3. Performance of the Proposed Method

Our goal is to make USV automatically navigate in the unknown environment and learn the water region during exploring. Figure 6 shows the performance of our online method during navigation. In the beginning, our segmentation network does not have any prior-knowledge of the water region. It cannot properly recognize the water surface. The segmentation performance is continuously improved as online training progresses, as shown in the first row of Figure 6. After sailing for a while, our system is adapted to the unknown environment and could recognize the water surface region. The second row of Figure 6 shows the segmentation results of images in different perspectives after enough training.
The performance of this approach can be better understood by the comparison with the state-of-the-art methods for semantic video segmentation. The first set of baseline methods, K-mean and graph-based segmentation method, are two representative conventional methods, which segment the image according to the feature of the pixel value and region edge. The second set of methods, UNet [26], RefineNet [31] and DeepLab [32], are Convolutional Neural Networks that need training with large labeled data.
Quantitative results are provided in Table 2. The segmentation quality is measured using average segmentation precision. In each frame, we measure the intersection divided by the union of the segmentation result and the ground truth labeled manually. Then, we average those scores over all frames in the sequence. Five videos of about 20–30 min in length recorded at different times and from different scenes are used to compare the performance of these algorithms. Each video is converted to images and about ten thousand images are selected as the experiment data. Twenty percent of the images are labeled as the test data in each video. For the traditional method, an image is segmented into different region according to the pixel value for the k-mean method and the image edge information for the graph-based method. The region connected to the free region in front of the USV is regarded as the water surface region. Segmentation precision is measured according to the corresponding test label. For the methods based on deep learning, eighty percent of the images are labeled as the training data in each video. Then these convolutional Neural Networks are trained with the training label data and tested by the test data. With our method, no pre-labeled training data are needed and learning is automatically. After the training is completed by itself, the test data is used to test our method.
As the traditional method is susceptible to the factors, such as waves, reflections, and illumination, the K-mean and graph-based methods have a poor performance. With enough labeled training data, the UNet, RefineNet and DeepLab perform well with high precision. However, the performances of these methods heavily depend on the training data set. Due to the complexity and variability of the navigation environment, only limited scenes are sampled and labeled as the training data. Without any pre-labeled training data, the performance of the proposed method is close to the state-of-the-art. With the continuous online learning, the proposed method is robust in the rich and varied scene. This can be manifested in testing case video3 and video5 where the proposed method shows better performance compared with UNet.

4. Conclusions

In this paper, an online semi-supervised learning method for water region segmentation is proposed for USVs in a dynamic navigation environment. Automatically labeled training data are generated by an adaptive multistage segmentation method with the aid of lidar data. An encoder-decoder semantic segmentation network is designed as the main work horse for the task. A modified binary cross entropy loss function with a weight map was applied to address the reliability issue in different pixels during the network training. Our experimental results showed that the proposed method outperformed the conventional methods. While having a similar performance as the state-of-art deep learning based algorithms, the proposed method has the advantage of self-training without the need of manually labeled data. With incremental and continuous learning online, the proposed method is robust and adaptive to the changing unknown environment.
In this paper, we verified the validity of the proposed method. However, we did not address some very hard situation for vision navigation, for example, water surface mirror reflection near shore. In some extreme cases, even a human eye can barely tell the difference between the shore objects and its mirror reflections on the water surface. Future work may direct to attack those difficult problems by multi-sensor data fusion and time-sequence deep learning algorithm such as recursive network (RNN).

Author Contributions

W.Z.: Implementation of algorithm and experiment analysis and writing of the paper. C.X.: Conceived the research project and proposed the main idea of the paper. Q.L.: Conceived the research project and proposed the main idea of the paper. Y.W.: Scientific advising and writing. C.Z.: Scientific advising and writing. H.Y.: Prepared the materials and assisted in completing the experiment. S.X.: Prepared the materials and assisted in completing the experiment. Y.Z.: Revised the paper. X.Z.: Revised the paper. X.L.: Revised the paper. All authors read and approved the final manuscript.

Funding

This research was funded by the National Science Foundation of China (grant number 51579204), Wuhan University of Technology Independent Innovation Research Foundation of China (grant number 2018III059GX), Hubei Provincial Natural Science Foundation of China through (number 2016CFB362), and Double First-rate Project of WUT. Q. Li acknowledges the support of Virginia Microelectronics Consortium (VMEC) Chair professor grant.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Matos, A.; Martins, A.; Dias, A.; Ferreira, B.; Almeida, J.M.; Ferreira, H.; Amaral, G.; Figueiredo, A.; Almeida, R.; Silva, F. Multiple robot operations for maritime search and rescue in euRathlon 2015 competition. In Proceedings of the OCEANS 2016-Shanghai, Shanghai, China, 10–13 April 2016; pp. 1–7. [Google Scholar]
  2. Yan, R.-J.; Pang, S.; Sun, H.-B.; Pang, Y.-J. Development and missions of unmanned surface vehicle. J. Mar. Sci. Appl. 2010, 9, 451–457. [Google Scholar] [CrossRef]
  3. Singh, Y.; Sharma, S.; Hatton, D.; Sutton, R. Optimal path planning of unmanned surface vehicles. Indian J. Geo-Mar. Sci. 2018, 47, 1325–1334. [Google Scholar]
  4. Singh, Y.; Sharma, S.; Sutton, R.; Hatton, D.; Khan, A. A constrained A* approach towards optimal path planning for an unmanned surface vehicle in a maritime environment containing dynamic obstacles and ocean currents. Ocean Eng. 2018, 169, 187–201. [Google Scholar] [CrossRef]
  5. Praczyk, T. Detection of land in marine images. Int. J. Comput. Intell. Syst. 2018, 12, 273–281. [Google Scholar] [CrossRef]
  6. Praczyk, T. A quick algorithm for horizon line detection in marine images. J. Mar. Sci. Technol. 2018, 23, 164–177. [Google Scholar] [CrossRef]
  7. Gershikov, E. Is color important for horizon line detection? In Proceedings of the 2014 International Conference on Advanced Technologies for Communications (ATC 2014), Hanoi, Vietnam, 15–17 October 2014; pp. 262–267. [Google Scholar]
  8. Gershikov, E.; Libe, T.; Kosolapov, S. Horizon line detection in marine images: which method to choose? Int. J. Adv. Intell. Syst. 2013, 6, 2013. [Google Scholar]
  9. Gershikov, E.; Baskin, C. Efficient horizon line detection using an energy function. In Proceedings of the International Conference on Research in Adaptive and Convergent Systems, Krakow, Poland, 20–23 September 2017; pp. 110–115. [Google Scholar]
  10. Zhan, W.-Q.; Xiao, C.-S.; Yuan, H.-W.; Wen, Y.-Q. Effective waterline detection for unmanned surface vehicles in inland water. In Proceedings of the 2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA), Montreal, QC, Canada, 28 November–1 December 2017; pp. 1–6. [Google Scholar]
  11. Ji, Z.; Su, Y.-T.; Wang, J.; Hua, R. Robust sea-sky-line detection based on horizontal projection and hough transformation. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009; pp. 1–4. [Google Scholar]
  12. Zafarifar, B.; Weda, H. Horizon detection based on sky-color and edge features. Proc. SPIE 2008, 6822. [Google Scholar] [CrossRef]
  13. Shen, Y.-F.; Krusienski, D.; Li, J.; Rahman, Z.-U. A hierarchical horizon detection algorithm. IEEE Geosci. Remote Sens. Lett. 2013, 10, 111–114. [Google Scholar] [CrossRef]
  14. Lu, J.-W.; Dong, Y.-Z.; Yuan, X.-H.; Lu, F.-L. An algorithm for locating sky-sea line. In Proceedings of the IEEE International Conference on Automation Science and Engineering, CASE’06, Shanghai, China, 8–10 October 2006; pp. 615–619. [Google Scholar]
  15. Yan, Y.; Shin, B.-S.; Mou, X.; Mou, W.; Wang, H. Efficient horizon detection on complex sea for sea surveillance. Int. J. Electr. Electron. Data Commun. 2015, 3, 49–52. [Google Scholar]
  16. Wang, H.; Wei, Z.; Wang, S.-S.; Ow, C.-S.; Ho, K.-T.; Feng, B. A vision-based obstacle detection system for unmanned surface vehicle. In Proceedings of the 2011 IEEE Conference on Robotics, Automation and Mechatronics (RAM), Qingdao, China, 17–19 September 2011; pp. 364–369. [Google Scholar]
  17. Ji, Z.; Yu, X.-J.; Yu, Y.-L.; He, Y.-Q. Manifold embedding for zero-shot recognition. Cogn. Syst. Res. 2019, 55, 34–43. [Google Scholar] [CrossRef]
  18. Ji, Z.; Xiong, K.-L.; Pang, Y.-W.; Li, X.-L. Video summarization with attention-based encoder-decoder networks. IEEE Trans. Circuits Syst. Video Technol. 2019. [Google Scholar] [CrossRef]
  19. Ji, Z.; Zhang, Y.-Y.; Pang, Y.-W.; Li, X.-L. Hypergraph dominant set based multi-video summarization. Signal Process. 2018, 148, 114–123. [Google Scholar] [CrossRef]
  20. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  21. Praczyk, T. Neural anti-collision system for autonomous surface vehicle. Neurocomputing 2015, 149, 559–572. [Google Scholar] [CrossRef]
  22. Praczyk, T. Artifcial neural networks application in maritime, coastal, spare positioning system. Theor. Appl. Inf. 2006, 18, 1175–1189. [Google Scholar]
  23. Fefilatyev, S.; Smarodzinava, V.; Hall, L.O.; Goldgof, D.B. Horizon detection using machine learning techniques. In Proceedings of the 2006 5th International Conference on Machine Learning and Applications (ICMLA’06), Orlando, FL, USA, 14–16 December 2006; pp. 17–21. [Google Scholar]
  24. Cheng, D.-C.; Meng, G.-F.; Cheng, G.-L.; Pan, C.-H. Senet: Structured edge network for sea–land segmentation. IEEE Geosci. Remote Sens. Lett. 2017, 14, 247–251. [Google Scholar] [CrossRef]
  25. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient graph-based image segmentation. Int. J. Comput. Vision 2004, 59, 167–181. [Google Scholar] [CrossRef]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  27. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  28. Krähenbühl, P.; Koltun, V. Efficient inference in fully connected CRFs with Gaussian edge potentials. Adv. Neural Inf. Process. Syst. 2011, 24, 109–117. [Google Scholar]
  29. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Tran. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
  30. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected CRFs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  31. Lin, G.; Milan, A.; Shen, C.; Reid, I. Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 1925–1934. [Google Scholar]
  32. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Tran. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The diagram of the proposed method.
Figure 1. The diagram of the proposed method.
Sensors 19 02216 g001
Figure 2. Segmentation network.
Figure 2. Segmentation network.
Sensors 19 02216 g002
Figure 3. The experimental USV.
Figure 3. The experimental USV.
Sensors 19 02216 g003
Figure 4. Adaptive segmentation result comparision.
Figure 4. Adaptive segmentation result comparision.
Sensors 19 02216 g004
Figure 5. Water region prediction and weight map generating.
Figure 5. Water region prediction and weight map generating.
Sensors 19 02216 g005
Figure 6. Segmentation performance during the online learning.
Figure 6. Segmentation performance during the online learning.
Sensors 19 02216 g006
Table 1. Parameter of the experimental USV and computer.
Table 1. Parameter of the experimental USV and computer.
EquipmentParameterValue/Model
USVlength3.96 m
width1.55 m
draft0.3~0.5 m
Max speed2.2 m/s
ComputerCPUIntel i7-5820
GPUTitan X
Memory32 GB
Table 2. Precision of the state-of-the-art methods and the proposed method.
Table 2. Precision of the state-of-the-art methods and the proposed method.
MethodVideo1Video2Video3Video4Video5
K-mean 160.165.259.562.767.2
Graph-based 173.664.556.763.571.4
UNet 297.995.396.297.195.8
RefineNet 29.597.698.197.998.4
DeepLab 299.79.578.799.298.7
Ours 397.394.397.296.496.3
1 The traditional feature-based method; 2 Methods based on deep learning with manual label training data set; 3 The proposed method that is self-learning with no manual label data set.

Share and Cite

MDPI and ACS Style

Zhan, W.; Xiao, C.; Wen, Y.; Zhou, C.; Yuan, H.; Xiu, S.; Zhang, Y.; Zou, X.; Liu, X.; Li, Q. Autonomous Visual Perception for Unmanned Surface Vehicle Navigation in an Unknown Environment. Sensors 2019, 19, 2216. https://doi.org/10.3390/s19102216

AMA Style

Zhan W, Xiao C, Wen Y, Zhou C, Yuan H, Xiu S, Zhang Y, Zou X, Liu X, Li Q. Autonomous Visual Perception for Unmanned Surface Vehicle Navigation in an Unknown Environment. Sensors. 2019; 19(10):2216. https://doi.org/10.3390/s19102216

Chicago/Turabian Style

Zhan, Wenqiang, Changshi Xiao, Yuanqiao Wen, Chunhui Zhou, Haiwen Yuan, Supu Xiu, Yimeng Zhang, Xiong Zou, Xin Liu, and Qiliang Li. 2019. "Autonomous Visual Perception for Unmanned Surface Vehicle Navigation in an Unknown Environment" Sensors 19, no. 10: 2216. https://doi.org/10.3390/s19102216

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop