Next Article in Journal
A Systematic Review of Industrial Exoskeletons for Injury Prevention: Efficacy Evaluation Metrics, Target Tasks, and Supported Body Postures
Previous Article in Journal
Pose2Sim: An End-to-End Workflow for 3D Markerless Sports Kinematics—Part 2: Accuracy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ship Segmentation and Georeferencing from Static Oblique View Images

German Aerospace Center (DLR), Institute for the Protection of Maritime Infrastructures, Fischkai 1, 27572 Bremerhaven, Germany
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2713; https://doi.org/10.3390/s22072713
Submission received: 17 February 2022 / Revised: 16 March 2022 / Accepted: 30 March 2022 / Published: 1 April 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Camera systems support the rapid assessment of ship traffic at ports, allowing for a better perspective of the maritime situation. However, optimal ship monitoring requires a level of automation that allows personnel to keep track of relevant variables in the maritime situation in an understandable and visualisable format. It therefore becomes important to have real-time recognition of ships present at the infrastructure, with their class and geographic position presented to the maritime situational awareness operator. This work presents a novel dataset, ShipSG, for the segmentation and georeferencing of ships in maritime monitoring scenes with a static oblique view. Moreover, an exploration of four instance segmentation methods, with a focus on robust (Mask-RCNN, DetectoRS) and real-time performances (YOLACT, Centermask-Lite) and their generalisation to other existing maritime datasets, is shown. Lastly, a method for georeferencing ship masks is proposed. This includes an automatic calculation of the pixel of the segmented ship to be georeferenced and the use of a homography to transform this pixel to geographic coordinates. DetectoRS provided the highest ship segmentation mAP of 0.747. The fastest segmentation method was Centermask-Lite, with 40.96 FPS. The accuracy of our georeferencing method was (22 ± 10) m for ships detected within a 400 m range, and (53 ± 24) m for ships over 400 m away from the camera.

1. Introduction

1.1. Maritime Situational Awareness and Ship Monitoring

Research in the field of maritime safety and security concentrates on the development, testing and validation of innovative systems for the assessment of the status of maritime infrastructures. One aspect of this is in the development of maritime situational awareness systems to quantitatively determine the protection status of infrastructures in real-time and execute measures to respond to threats (e.g., major accidents, natural catastrophes, terror attacks, organised crime) [1]. An automatic and meaningful awareness of the maritime situation requires instruments and sensors that are able to recognise elements of interest to propose suitable measures to the user and authorities [2].
Real-time ship monitoring at ports is one of today’s most challenging tasks for Vessel Traffic Services (VTS) [3], and great efforts are being made using Automatic Identification Systems (AIS) [4]. The International Maritime Organisation (https://www.imo.org/en/OurWork/Safety/Pages/AIS.aspx, accessed on 16 February 2022) requires that every ship and vessel with 300 or more gross tonnage carries AIS transceivers, which must transmit information such as a unique identification number, ship type, position, course and speed, in the form of encoded radio messages. The AIS tracking system allows VTS and the surrounding ships to be aware of the marine traffic in the area and to perform tasks such as collision avoidance or search and rescue. However, the AIS messages are only transmitted by ships in intervals from 2 to 10 s while underway, and up to 6 min in static position (https://www.navcen.uscg.gov/?pageName=AISMessages, accessed on 16 February 2022). This leaves room for real-time systems to analyse the situation several times per second and therefore help to prevent or respond to complications. Moreover, since it is an open standard, AIS presents some vulnerabilities to cyber threats, which include spoofing, highjacking attacks or denial of service [5,6,7,8,9].
Optical camera systems thus become a good choice for the rapid assesment of ship traffic. Due to their availability, affordable price and uncomplicated deployment, they can serve as support in tasks of ship monitoring. When placed with elevation and an oblique view of the water, they allow an excellent perspective of the situation. Personnel might, however, not be able to effectively keep track of everything relevant due to the amount of video screens and information that need to be assessed [10]. It therefore becomes important to have automatic recognition of ships present at the infrastructure, with their class and position presented to the maritime situational awareness operator in a more understandable format, for example on a map.

1.2. Ship Segmentation

Deep-learning-based object recognition is one of most well studied computer vision topics and can be applied to recognise ships in images from optical cameras.
One of the trends in the field is to obtain the mask of objects detected with segmentation, which is referred to by the name of instance segmentation. This technique allows the extraction of further information relating to the recognised objects, such as their position on Earth’s coordinate system, referred to as georeferencing. Georeferencing can be better inferred from the segmented mask of an object than from a sourrounding bounding box, which usually contains a lot of unnecessary background.
Instance segmentation is a type of supervised deep learning problem, and datasets for training are needed. There are some general-purpose benchmark datasets, such as COCO [11]. The images available in the general-purpose datasets do not suit specific tasks of ship recognition with the precision required by a maritime awareness application. Real-world maritime situations require varied ship data with precise annotations, that should include the ship class, mask and further features, such as geographic coordinates. Amongst works in the literature that deal with ship detection with optical monitoring views, available datasets are the Singapore Maritime Dataset (SMD) [12], Seaships7000 [13] and the dataset presented by Chen et al. [14]. However, these datasets lack the required annotations to perform instance segmentation, contain low ship variety, and do not have a favourable oblique view, which means that georeferencing becomes difficult to perform. We have not found a public dataset that contains oblique view images of a maritime infrastructure and mask annotations of several types of ships for the exploration, evaluation and development of instance segmentation methods.
State-of-the-art instance segmentation methods with robust results on the COCO dataset are Mask-RCNN [15] and DetectoRS [16]. State-of-the-art methods for real-time applications have been developed in YOLACT [17] and Centermask [18]. Existing works evaluate object detection methods in maritime environments on their private ship detection datasets [19,20]. Nita et al., in [21], tackle the task of ship instance segmentation without a real-time approach, using only Mask-RCNN [15] on their private dataset. A comparison of state-of-the-art ship instance segmentation methods from maritime oblique view images, with a focus on robust and real-time methods, and a public dataset, has not been found.

1.3. Ship Georeferencing

Once ships have been detected and segmented, georeferencing is required to provide their real-time location to a situational awareness system.
General-purpose object georeferencing has been studied primarily for airborne applications [22,23,24] and autonomous driving [25]. In the field of ship georeferencing, existing works made use of radar [26], remote sensing using synthetic aperture radar [27] and AIS [28]. Helgesen et al., in [29], proposed a pipeline for ship detection and georeferencing using their private dataset with oblique view images of the water. They use the pinhole camera calibration transformation matrix proposed in [30] to georeference bounding boxes of detected ships. For this, their approach requires previous knowledge of the camera, such as location, elevation, field of view and tilt angle of the lense. Moreover, their study is limited to a maximum range of 400 m between the camera and the ship. An extensive analysis of a methodology for ship georeferencing using oblique view images where previous knowledge of the camera is not required has not been found.

1.4. Proposed Work

The aim of this work is to address the gaps in the field of ship segmentation and georeferencing from oblique view images to advance the development of maritime situational awareness systems with higher levels of automated information extraction. Our proposed contributions can be summed up in the following three statements.
First, we present the creation of a novel dataset, ShipSG, for ship segmentation and georeferencing using static oblique view images. This dataset contains mask annotations of the ships present in the images along with their corresponding class, positions (latitude, longitude), and lengths. The dataset was created using two cameras at a port location, and the geographic ship positions were obtained using AIS data. To the best of our knowledge, this is the first dataset of its kind and will be publicly available (https://dlr.de/mi/shipsg, accessed on 16 February 2022).
Second, we explore four instance segmentation methods to recognise ships with the ShipSG dataset. Two as a baseline for robust instance segmentation, Mask-RCNN [15] and DetectoRS [16], and two capable of real-time processing, YOLACT [17] and Centermask-Lite [18]. The goal is to find which of the four provides the best precision for ship segmentation, and which method provides the best trade-off between precision and inference speed. We also provide an approximation of how well these methods generalise after training with the ShipSG dataset on the aforementioned datasets SMD [12], Seaships7000 [13] and the dataset by Chen et al. [14].
Third, we propose a methodology for the automatic georeferencing of ship masks. We automatically calculate the pixel to be geoferenced from the segmented masks provided by the previous instance segmentation step. The georeferencing method we propose is based on the use of a homography matrix to transform pixels from the ShipSG images to geographic latitudes and longitudes.
The following sections of this paper are organised as follows. Section 2 presents the creation of the ShipSG dataset and its content. Also discussed are the selected instance segmentation methods and our proposed ship georeferencing method. Section 3 shows the results of each instance segmentation method along with an analysis of our ship georeferencing method. We present the results in Section 4, followed by our conclusion in Section 5.

2. Materials and Methods

2.1. The ShipSG Dataset

The ShipSG dataset (https://dlr.de/mi/shipsg, accessed on 16 February 2022) was collected using two cameras located at the Fischereihafen-Doppelschleuse lock, part of the port of Bremerhaven, Germany. The cameras have partly overlapping fields of view, facing the port basin, in order to observe the maritime activity at the entrance of the lock and at the river where the port is located (see Figure 1). The port basin is within a 400 m range of the cameras. The range of the river is over 400 m, where ships can be seen up to approximately 1200 m away from the cameras. The acquisition of images took place in Autumn 2020 during daylight hours with sunny, cloudy, windy and rainy weather conditions. The tidal range was between 3 and 4 m (https://gezeitenfisch.com/de/bremen/bremerhaven-doppelschleuse, accessed on 16 February 2022). Vehicles and people appearing within the images were anonymised since they are not of interest for this work.
To obtain information relating to ships present within each image, we accessed AIS position and static messages (https://www.navcen.uscg.gov/?pageName=AISMessages, accessed on 16 February 2022) from AISHub (https://www.aishub.net/, accessed on 16 February 2022). The former is sent by ships in intervals between 2 and 10 s, and the latter in intervals of 3 and 6 min [4]. These messages contain, amongst other fields, the ship position (latitude and longitude in decimal degrees) and the ship length (in meters). We used these two fields to annotate the ships in the images of the dataset. The AIS ship position is used as a ground truth for our georeferencing method and the ship length is used to study how our georeferencing method changes with the ship length. In order to label data, we accessed the timestamp of each image and searched for the AIS message which has the most similar timestamp and a position which lies within the field of view of the cameras. We defined 100 milliseconds as the maximum offset between the image and AIS timestamp so that the position of the ship seen in the image corresponds as close as possible with the position contained in the AIS message. Since a short offset is used, and due to the fact that ships send AIS messages with a time period of seconds, this leads to only one AIS reference of a ship per image. We discarded images in which the timestamp could not be related to any AIS message timestamp. A total of 3505 images were found with an AIS message corresponding to one of the ships within the image.
We designated the ship classes for the dataset based on an observation of their purpose and visual similarities. Examples of each ship class are shown in Figure 2 and are described as follows:
  • Cargo: All types of cargo ships.
  • Law Enforcement: Police watercrafts and coast guard ships.
  • Passenger/Pleasure: Ferries, yachts, pleasure and sailing crafts.
  • Special 1: Crane vessels, dredgers and fishing boats.
  • Special 2: Research and survey ships, search and rescue ships and pilot vessels.
  • Tanker: All types of tankers.
  • Tug: All types of tugboats.
For the task of instance segmentation, annotations of ship masks are needed as input for algorithm training. We manually annotated the ship masks within each image with their corresponding class using the LabelMe software [31]. Figure 3 illustrates samples of our dataset with the annotated masks.
We used the definition of small, medium and large mask area scales that were introduced for the COCO dataset [11], and have the following values:
M a s k A r e a S c a l e = Small , if a r e a 32 2 pixels , Medium , if 32 2 < a r e a 96 2 pixels , Large , if a r e a > 96 2 pixels
These area scales are later used in this work to measure the performance of instance segmentation algorithms. Table 1 lists the number of masks annotated in the dataset per class, as well as the number of masks per area scale in pixels. In total, 11625 masks were annotated.
In summary, the ShipSG dataset contains:
  • 3505 images from the two cameras.
  • 11,625 annotated ship masks grouped in seven classes with COCO format [11]. The AIS ship type will also be shared so that future users can create their own classes.
  • 3505 geographic positions, consisting of the latitude and longitude of one of the masks within each image.
  • 3505 ship lengths, one per geographic position annotated.
The authors of this paper intend to provide the dataset to the public (https://dlr.de/mi/shipsg, accessed on 16 February 2022). To the best of our knowledge, this dataset is the first of its kind dedicated to ship segmentation and georeferencing, and which is available to the public. It was not possible to share the MMSIs associated with the AIS messages for the ShipSG dataset due to the underlying principles of the privacy policy implemented when performing data acquisition. This policy was composed by the German Aerospace Center (DLR) and implements the General Data Protection Regulation (EU) 2016/679 [32].

2.2. Instance Segmentation Methods Selected

In order to explore the task of instance segmentation using our dataset, four state-of-the-art methods were selected. Two as a baseline for robust instance segmentation and two capable of real-time processing, which are described in Section 2.2.1 and Section 2.2.2, respectively. The comparison of these methods follows the standard metrics used by the COCO dataset [11], based on the well-known average precision (AP). The mean average precision (mAP) is the calculated mean of all the AP of the classes present. As a metric of speed during inference, frames per second (FPS) are considered.

2.2.1. Robust Instance Segmentation Methods

Mask R-CNN [15] is a two-stage algorithm that was developed as an extention of the object detector Faster R-CNN [33]. In the first stage, with the region proposal network [33], multiple object candidates are proposed. In the second stage, the region of interest pooling extracts features from each candidate and performs the classification of the object and the regression of the bounding box. In Mask R-CNN, a fully convolutional network was added to regress the mask from the detected bounding boxes. This method is one of the most popular in the field of instance segmentation for its robustness. With the ResNeXt-101 backbone [34], it achieves a mask mAP of 0.375 on the COCO dataset.
DetectoRS [16] is a multi-stage network that proposed a recursive feature pyramid [35] to include additional feedback connections from feature pyramid networks into the bottom-up backbone layers. Its authors also propose the convolution of features by looking twice at the input with different atrous rates and then to combine the outputs, which is referred to as switchable atrous convolution. This method is a state-of-the-art instance segmentation method, and with ResNet-50 as the backbone [36], it achieves a mask mAP of 0.444 on the COCO dataset.

2.2.2. Real-Time Instance Segmentation Methods

YOLACT [17] emerged as one of the first real-time and one-stage instance segmentation methods. It uses an independent fully convolutional network [37] to produce prototype masks and a parallel branch to calculate mask coefficients for each predicted anchor box, which are filtered using non-maximum supression. Both branches are combined by cropping and thresholding the prototyped masks with the filtered anchor box. On the COCO dataset, with ResNet-101 as the backbone [36] and 700 × 700 pixels as an input size, it achieves an mAP of 0.312 and an inference speed of 23.4 FPS.
Centermask [18] is a one-stage method. It makes use of the fully convolutional one-stage object detector [38], and introduces a spatial attention-guided mask branch, which is paired with the object detector to suppress the pixels that do not belong to the mask on the regions proposed as boxes. Specifically, for our work, we selected Centermask-Lite, a downsized version of the original which is better suited for real-time applications. The authors of Centermask introduced in [39] a novel backbone, VoVNet, where instead of adding residual shortcuts every second feature map, as is done in ResNet, features are concatenated only once in the last feature map. With VoVNet-39 as backbone, Centermask-Lite achieves a mask mAP of 0.363 and 35.7 FPS on the COCO dataset.

2.3. Ship Georeferencing Using Homography

Once the ships are segmented using an instance segmentation method, georeferencing is required to provide their location to the situational awareness system.
The use of homography, an isomorphism between projective spaces, is well established in the computer vision field to transform points from the same planar surface captured by two perspectives, which up to now has not been deeply studied in the context of ship georeferencing.
We tested the use of a homography qualitatively for ship georeferencing for maritime anomaly detection environments in [40]. The present work will expand upon this and make an in depth quantitative study of the use of homographies for ship georeferencing.
We take advantage of the static view of the cameras and perform a transformation by calculating the homography between the camera pixel coordinates ( C x , C y ) and Earth’s geographic latitude and longitude ( φ , λ ) in decimal degrees. Since the ground surface area captured by the cameras is small enough with respect to the Earth’s local curvature, it is approximated by its tangent plane.
A homography, expressed as the matrix H, used to perform the transformation is shown in Equation (1), where the unknown parameters h 11 , h 12 , , h 32 are calculated using n number of pixel pairs ( C x , C y ) and geographic latitude and longitude pairs ( φ , λ ) as shown in Equation (2). Using n = 4 pairs of pixel coordinates and their corresponding geographic positions would suffice to solve the eight unkown parameters of Equation (2) ( h 11 , h 12 , , h 32 ). Having more pairs would result in more than one solution for each parameter of the equation. Therefore, n > 4 is preferred, allowing the optimal solution for each unknown parameter to be calculated using least squares.
φ λ 1 = H C x C y 1 = h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 1 C x C y 1 .
C x 1 C y 1 1 0 0 0 φ 1 · C x 1 φ 1 · C y 1 0 0 0 C x 1 C y 1 1 λ 1 · C x 1 λ 1 · C y 1 C x 2 C y 2 1 0 0 0 φ 2 · C x 2 φ 2 · C y 2 0 0 0 C x 2 C y 2 1 λ 2 · C x 2 λ 2 · C y 2 C x n C y n 1 0 0 0 φ n · C x n φ n · C y n 0 0 0 C x n C y n 1 λ n · C x n λ n · C y n h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 = φ 1 λ 1 φ 2 λ 2 φ n λ n
The n pairs have to be selected manually in advance to create H. Multiple pairs distributed throughout the geographic and image area of interest should be selected, for a better adjusted H. For ( φ 1 n , λ 1 n ) we used the geographic annotation that each image contains, as explained in Section 2.1, and for ( C x 1 n , C y 1 n ) the pixel coordinates of the corresponding ship in the corresponding image. To manually select this pixel, we observed the placement of the navigation antenna on each ship, since this is the element which provides the geographic location in the AIS messages. This antenna is usually located on the bridge or wheelhouse of the ship. We therefore selected the pixel that intersects the ship hull at the antenna location and the water underneath as the pixel corresponding to the latitude and longitude gathered with AIS.
We took n = 200 samples of the training set images to create the homographies of both cameras (see Figure 4), and solved Equation (2) to obtain H. The validation set is later used to quantitatively analyse how well this method performs.The separation of homographies by high or low tides was not found to provide a significant improvement in results.
Once the homographies are created, we then propose a method to automatically calculate the pixel ( C x , C y ) of the mask which best represents the ship’s geographic position. This pixel is afterwards georeferenced using Equation (1). We automatically find the pixel which lies at the intersection point between the ship hull and the water below the bridge or wheelhouse, where the navigation antenna of the ship is located. We calculate this pixel to be the lowest of the mask in the vertical direction (Y) corresponding to the statistical mode of the horizontal axis (X). An example of this procedure can be seen in Figure 5.

3. Results

3.1. Experimental Evaluation of Instance Segmentation Methods on the Dataset

We split the dataset into two sets of images—training and validation. The training set contains 80% of the dataset, with 2804 images, and the remaining 20% is used for validation, with 701 images. Both sets have a comparable class distribution, as the one shown in the last column of Table 1.
The training and evaluation of the robust methods discussed in Section 2.2.1, Mask R-CNN and DetectoRS, was done using the MMdetection framework from OpenMMLab [41]. The input image size in both cases was 1333 × 800 pixels, and the backbones selected were ResNeXt-101 and ResNet-50, respectively.
For the experimental setup of the real-time methods discussed in Section 2.2.2, our interest was to look for the optimal trade-off between inference speed and AP. Therefore we selected two configurations for each, one with a deeper backbone (more layers) and one with a shallower backbone. The implementation of YOLACT used was the source provided by its authors [17]. The first configuration, YOLACT550, uses a smaller input size (550 × 550 pixels) and the lighter ResNet-50 [36] as backbone and which will provide faster inference speed. The second configuration, YOLACT700, uses a higher input size of 700 × 700 pixels and a deeper backbone, the ResNet-101 [36], which will provide higher AP. As for Centermask-Lite, we used the source implementation provided by its authors [18], which is implemented on top of the framework Detectron2 [42]. For a faster inference speed, the first configuration, Centermask-Litev19, uses a lighter backbone, VoVNet-19 [39]. The second, Centermask-Litev39, uses the deeper VoVNet-39 [39], for higher AP. Both Centermask-Lite configurations use an input size of 800 × 600 pixels, as per definition by their authors [18].
Table 2 shows a summary of the training configurations. All the methods were initialised with weights pre-trained on the COCO dataset [11]. The Pytorch version used for the four methods was 1.8.1 and the GPU used to train and compute the inference speed was a Nvidia Quadro GV100.
Table 3 illustrates the results per method evaluated. For the evaluation and comparison of these methods, we chose the standard metrics used by the COCO dataset [11]. These are the overall mAP, the mAP at intersection over union 50% (mAP50), the mAP at intersection over union 75% (mAP75) and the mAP at different mask area scale, i.e., mAPS for small objects, mAPM for medium objects and mAPL for large objects (https://cocodataset.org/#detection-eval, accessed on 16 February 2022). As well as the mAP, we include the class-agnostic mask AP (APca) which shows that there is not significant class imbalance. As a metric of speed during inference, frames per second (FPS) are considered. The complete results showing each AP per class and per instance segmentation method are shown in Appendix A.
Out of the robust methods evaluated, Mask R-CNN [15] and DetectoRS [16] reach the best mask mAP in all cases compared to the real-time methods. DetectoRS achieves the highest APca and mAP, with values of 0.780 and 0.747, respectively. The inference speed of these methods provides 8.50 and 6.62 FPS, respectively.
Comparing the performances of the real-time methods, YOLACT [17] and Centermask-Lite [18], we observe that Centermask-LiteV39 achieves the highest APca and mAP. Moreover, Centermask-Lite in both configurations performs better with small and medium objects. YOLACT, however, achieves a higher mAP with large objects, 0.751, compared with Centermask-LiteV39, 0.661. In terms of inference speed, Centermask-LiteV19 achieves the highest FPS value of 40.98.

3.2. Generalisation of the Evaluated Instance Segmentation Methods on Other Datasets

It is advantageous for a computer vision architecture to be able to perform well in diverse scenarios, since this indicates that overfitting due to the dataset used has not ocurred.
To study the ability of the instance segmentation models trained in Section 3.1 to generalise in different scenarios, we analysed their performance with test images from other existing datasets with similar characteristics: SMD [12], Seaships7000 [13] and the dataset by Chen et al. [14]. Since these datasets do not provide ship mask annotations, we annotated the ships present in 100 images of the three datasets and combined them into a mini-dataset with a single class. Samples of this mini-dataset for generalisation can be seen in Figure 6. The annotated content is as follows:
  • SMD [12]: Two images per on-shore scene, totalling 80 images.
  • Seaships7000 [13]: 12 random images of the dataset.
  • Dataset by Chen et al. [14]: Two images per scene, totalling eight images.
Table 4 shows the results after inference on the generalisation mini-dataset. The best AP is achieved by DetectoRS with 0.486. When looking at the AP50, we observe that DetectoRS provides a satisfactory value of 0.830. Of the real-time methods, Centermask-LiteV39 achieves the best AP, with 0.387, and an AP50 of 0.710.These results show that the models trained with ShipSG can predict ships from other datasets with reasonable accuracy.

3.3. Experimental Evaluation of Ship Georeferencing

The data collection process, explained in Section 2.3, shows that every image of the dataset contains one ship position taken from AIS messages (latitude and longitude) along with the corresponding pixel annotation. We created two homographies, one per camera, as shown in Section 2.3. We then quantitatively analysed how well the method performs for the task of ship goereferencing.
We took the resulting ship masks from DetectoRS [16] using the 701 images of the validation set, which contains ship images from both cameras. We georeferenced the masks using Equation (1), after following the proposed method described in Section 2.3.
Figure 7 shows the qualitative results of our proposed ship georeferencing method. As introduced in Section 2.1, we split the field of view into the port basin area and the river area in order to observe the results from both camera ranges to the ships. These are smaller and greater than 400 m, respectively.
We quantitatively compared the true latitudes and longitudes collected with AIS ( φ A I S , λ A I S ) and the homography-georeferenced latitudes and longitudes ( φ H , λ H ). For this comparison, we convert both latitudes and longitudes from decimal degrees to Universal Transverse Mercator (UTM) to express every result in meters. The metrics used to determine how well the technique performs are the following:
  • Latitude absolute error ( Δ φ ):
    Δ φ = | φ A I S φ H |
  • Longitude absolute error ( Δ λ ):
    Δ λ = | λ A I S λ H |
  • Georeferencing distance error ( G D E ), to measure the distance between true and georeferenced positions. The haversine equation is used instead of euclidean distance, to take into account the radius (R) and therefore curvature of the Earth:
    G D E = 2 · R · arcsin sin 2 | φ A I S φ H | 2 + cos φ A I S · cos φ H · sin 2 | λ A I S λ H | 2
  • Distance root-mean-square error (DRMSE) [43], to measure the quadratic mean of all latitude and longitude errors. Due to the squared differences, larger errors are more penalised than small errors:
    R M S E φ = 1 k Σ i = 1 k ( φ A I S i φ H i ) 2
    R M S E λ = 1 k Σ i = 1 k ( λ A I S i λ H i ) 2
    D R M S E = R M S E φ 2 + R M S E λ 2
Table 5 shows the quantitative results. For all metrics, a lower value indicates a better result. As expected, our method is more accurate the closer the ships are to the cameras, which is represented by the smaller values of all metrics within the port basin. This can be compared to the river area, where every pixel covers more geographical area, and therefore the error becomes more significant. We consider the mean GDE as the most representative metric, because it directly measures the distance in meters between the actual and estimated positions. This metric reaches ( 22 ± 10 ) m inside the port basin and ( 53 ± 24 ) m on the river. The DRMSE calculated inside the port basin is 27 m, and 61 m on the river, providing a comparable result to the GDE metric. This indicates that errors larger than the mean GDEs are not common and do not have great impact, showing therefore that the method works consistently.
In Section 2.1 it was described that we collected ship lengths along with the ship positions from AIS messages. This was done to observe how the GDE changes with the ship length, as shown in Figure 8. For the smallest ship lengths (0 to 20 m), the GDE within the port basin and river are similar. This shows that, independently from the range between ship and camera, the smaller the ship, the more accurately the method finds the pixel of the mask to be georeferenced.
Range is the largest contributor to the GDE. For ships over 20 m long, there is a significant increase in GDE when viewed at ranges greater than 400 m (at the river). For ranges closer than 400 m (within the port basin), the ship length has less impact on the resulting GDEs than on the river.
Even though the larger and more distant the ship, the more difficult it becomes to find the exact pixel of the mask to be georeferenced, all GDEs are consistent within uncertainties per ship length. This shows that the method provides a reliable estimation, including for larger ships on the river side where every pixel covers more geographical area.

4. Discussion

In this work, ShipSG, a novel dataset for ship segmentation and georeferencing, has been presented. This dataset contains 3505 images from a static oblique view, using two cameras with partly overlapping views. A total of 11,625 ship masks were annotated and grouped in seven ship classes. Through the use of AIS messages, acquired simultaneously with the images, we annotated the geographic position (latitude and longitude) and length of one ship per image. To the best of our knowledge, this is the first dataset of its kind. This dataset will be shared with the public (https://dlr.de/mi/shipsg, accessed on 16 February 2022). We split the ShipSG dataset into training (80%) and validation (20%) sets, and explored four state-of-the-art instance segmentation methods for the automatic recognition of the annotated ships. Two robust methods, Mask-RCNN [15] and DetectoRS [16], and two real-time methods, YOLACT [17] and Centermask-Lite [18], were explored. Each real-time method was studied with two configurations, with a lighter and a deeper backbone. After training all methods with the ShipSG dataset, DetectoRS provides the best mAP, of 0.747. The fastest method explored was Centermask-LiteV19, with 40.96 FPS. Centermask-LiteV39, with the deeper backbone Vovnet-39 [39], provided the best trade-off between mAP and FPS, with 0.644 and 35.25, respectively, which makes it the most suitable candidate for tasks of real-time ship segmentation for a future situational awareness system. YOLACT700, however, with ResNet-101 as the backbone [36], performs better with large objects (mAPL of 0.751) when compared with Centermask-Lite (mAPL of 0.661). Since the key aim of this work is to recognise ships at all ranges and with all sizes and classes, future work will focus on improving large object segmentation for Centermask-Lite.
As all instance segmentation methods provided a good mAP, higher than 0.5 in all cases, we tested how well they generalise when segmenting ships of other maritime datasets after training with ShipSG. We annotated the ship masks of a mini-dataset of 100 images from SMD [12], Seaships7000 [13] and the dataset by Chen et al. [14], which are datasets that either did not contain the necessary annotations or are not suitable due to the lack of variety of ships and scenes, but still contain some ships that could be used for testing. DetectoRS still showed the best AP, with 0.486, and AP50 of 0.830. Centermask-LiteV39 provided the best AP of the real-time methods explored, with 0.387 and AP50 of 0.710. It has been shown, therefore, that the methods trained with the ShipSG dataset could be used for inference on other similar maritime scenes.
A method for the automatic geoferencing of ship masks has been presented. The method is based on the use of a homography matrix to transform pixels from the ShipSG images, taking advantage of the static view, to geographic latitudes and longitudes. The homographies, one per camera, were created using the AIS positions of the ships present in the training set images of ShipSG. The georeferenced pixel is chosen to be that which intersects the ship hull and the water below, at the point where the navigation antenna is located on the ship. We also present a method to automatically calculate this pixel.
We quantitatively analysed our proposed method for ship georeferencing at a range closer than 400 m (within the port basin) and farther than 400 m (on the river). As expected, the accuracy of the method is best when the ship is closest to the camera. Furthermore, independently from the range between the ship and camera, the smaller the ship, the more accurately the method finds the pixel of the mask to be georeferenced. The results prove that this is a reliable approach, since the georeferenced pixel is shown to fall within the bounds of the ship (Figure 8) and within the uncertainty of the method (Table 5) when considering both shorter and longer camera ranges.
A future maritime situational awareness tool for ship georeferencing used by, for instance, authorities, would need a series of further improvements to avoid the estimation of the presence of recognised ships in a location where an error could be more significant. This is, for instance, the case of an estimation of a ship position sitting on land. Future work will include the study and mitigation of the systematic effects of the use of homographies, to improve the accuracy of the method. Furthermore, a future approach will make use of deep learning for the identification of the georeferenced pixel from the masks to minimise the presented georeferencing results. The use of deep learning will include the manual annotation of the pixel to be georeferenced from all the masks of the dataset as ground truth. The manual annotation of the pixel to be georeferenced can also be used as a baseline to analyse how good humans are at defining the pixel of a ship to be georeferenced against an automatic approach like the one we propose in our work. Further considerations will also include the automatic calculation of other ship parameters such as ship length, since the corresponding ground truth values are already available within the dataset. Future improvements of the dataset will also be shared with the public.
Our georeferencing method offers several improvements when compared with the state of the art in ship georeferencing from static oblique view images [29]. Firstly, our methodology can be replicated using any existing static camera at a maritime infrastructure without previous knowledge about the camera, such as location, elevation, field of view and tilt angle. Moreover, our quantitative analysis includes seven classes of ships and many ship sizes, from small boats to large container ships. We also analysed results in two independent ranges, within and over 400 m.
The ship segmentation and gereferencing method presented in this work is intended to be utilised as part of a complete pipeline for ship segmentation and georeferencing that can be used to present meaningful real-time information about ships to maritime situational awareness operators.

5. Conclusions

A novel dataset, ShipSG, for ship segmentation and georeferencing using a static oblique view of a port has been presented. This dataset contains images with mask annotations of ships present, and their corresponding class, position and length.
Four instance segmentation methods to recognise ships were explored using the dataset. DetectoRS shows the best overall mAP, though Centermask-LiteV39 is found to be the most precise of the real-time capable methods studied and is therefore most suited for our application. After training with ShipSG, the generalisation on a mini-dataset made of existing maritime datasets is shown.
A quantitative analysis of our homography based method for ship georeferencing from their segmented masks has also been presented. As expected, the accuracy of the method is best when the ship is closest to the camera. However, results prove that this is a reliable approach for all ship lengths independently from the range, since the georeferenced pixel is shown to fall within the bounds of the ship when considering both shorter and longer camera ranges.
Future studies will focus on the improvement of Centermask-LiteV39 to detect larger objects, the study of systematic effects of homographies for georeferencing, the use of deep learning to improve the identification of the pixel from the mask to be georeferenced and the integration of both tasks on a single pipeline that can be used by a maritime situational awareness system.

Author Contributions

Conceptualisation, B.C.-P., S.B. and M.S.; methodology, B.C.-P.; software, B.C.-P.; validation, B.C.-P.; formal analysis, B.C.-P.; investigation, B.C.-P., S.B. and M.S.; data curation, B.C.-P.; writing—original draft preparation, B.C.-P.; writing—review and editing, B.C.-P., S.B. and M.S.; visualisation, B.C.-P.; supervision, S.B. and M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented and analysed in this work can be found here: https://dlr.de/mi/shipsg, accessed on 16 February 2022.

Acknowledgments

The authors would like to thank Jens-Michael Schlüter and Michael Busack from the Alfred Wegener Institute for Polar and Marine Research for their technical support.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
VTSVessel Traffic Services
AISAutomatic Identification System
SMDSingapore Maritime Dataset
DLRGerman Aerospace Center
APAverage Precision
mAPmean Average Precision
FPSFrames Per Second
UTMUniversal Transverse Mercator
GDEGeoreferencing Distance Error
DRMSEDistance Root-Mean-Square Error

Appendix A. Class Average Precision per Class and Instance Segmentation Method

This appendix contains the resulting instance segmentation APs per class and method explored in Section 2.2.
Table A1. Resulting instance segmentation APs per class of the ShipSG dataset using Mask-RCNN, as explained in Section 3.1.
Table A1. Resulting instance segmentation APs per class of the ShipSG dataset using Mask-RCNN, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.7070.9870.9380.5410.7150.762
Law Enforcement0.8420.990.9780.5580.8680.73
Passenger/Pleasure0.7050.9510.9190.5420.7370.72
Special 10.7540.9470.9230.3570.7220.83
Special 20.7730.9580.9270.510.8080.81
Tanker0.6770.9730.860.520.6850.757
Tug0.6710.9210.8510.4910.7280.797
Table A2. Resulting instance segmentation APs per class of the ShipSG dataset using DetectoRS, as explained in Section 3.1.
Table A2. Resulting instance segmentation APs per class of the ShipSG dataset using DetectoRS, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.7180.9950.9330.590.7110.788
Law Enforcement0.8480.990.9790.5630.8690.778
Passenger/Pleasure0.7330.9830.9520.620.7560.761
Special 10.7860.9750.9440.4660.7390.847
Special 20.7720.9850.9350.5840.7990.814
Tanker0.6870.9940.8450.5340.6950.767
Tug0.6850.9490.8820.5390.7280.791
Table A3. Resulting instance segmentation APs per class of the ShipSG dataset using YOLACT550, as explained in Section 3.1.
Table A3. Resulting instance segmentation APs per class of the ShipSG dataset using YOLACT550, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.4690.8890.4240.080.4380.722
Law Enforcement0.6850.9780.910.0720.7030.717
Passenger/Pleasure0.5120.8870.5890.110.5460.633
Special 10.6170.9060.7710.0390.4460.745
Special 20.510.9330.5520.1550.5450.716
Tanker0.4420.820.5040.0650.4540.668
Tug0.4540.7920.5110.0790.4710.766
Table A4. Resulting instance segmentation APs per class of the ShipSG dataset using YOLACT700, as explained in Section 3.1.
Table A4. Resulting instance segmentation APs per class of the ShipSG dataset using YOLACT700, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.5290.9110.540.130.5060.728
Law Enforcement0.7270.9780.9230.1230.7450.76
Passenger/Pleasure0.5770.8960.7680.1970.5950.728
Special 10.6690.9370.8090.0760.5290.782
Special 20.570.9470.7390.2030.6090.775
Tanker0.5080.9010.5590.1370.5360.704
Tug0.4940.8080.5580.1120.5510.78
Table A5. Resulting instance segmentation APs per class of the ShipSG dataset using Centermask-LiteV19, as explained in Section 3.1.
Table A5. Resulting instance segmentation APs per class of the ShipSG dataset using Centermask-LiteV19, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.5970.8480.7670.4970.5850.644
Law Enforcement0.7390.8530.8540.5200.7630.626
Passenger/Pleasure0.6080.8440.7780.4840.6270.625
Special 10.6610.8430.7980.3130.5710.719
Special 20.6650.8380.7980.4290.6860.674
Tanker0.5760.8320.7300.4380.5890.618
Tug0.6010.8240.7360.5040.6570.695
Table A6. Resulting instance segmentation APs per class of the ShipSG dataset using Centermask-LiteV39, as explained in Section 3.1.
Table A6. Resulting instance segmentation APs per class of the ShipSG dataset using Centermask-LiteV39, as explained in Section 3.1.
ClassAPAP50AP75APSAPMAPL
Cargo0.5980.8230.7840.4770.5940.638
Law Enforcement0.7310.8350.8430.4980.7560.612
Passenger/Pleasure0.6070.8250.7790.5170.6290.622
Special 10.6520.8180.7850.2690.5550.733
Special 20.6570.8260.7860.4360.6760.684
Tanker0.5650.8150.7320.4420.5780.616
Tug0.6990.930.7990.590.7520.721

References

  1. Engler, E.; Göge, D.; Brusch, S. ResilienceN—A multi-dimensional challenge for maritime infrastructures. NAŠE MORE Znanstveni časopis za More i Pomorstvo 2018, 65, 123–129. [Google Scholar]
  2. Wang, K.; Liang, M.; Li, Y.; Liu, J.; Liu, R.W. Maritime traffic data visualization: A brief review. In Proceedings of the 2019 IEEE 4th International Conference on Big Data Analytics (ICBDA), Suzhou, China, 15–18 March 2019; pp. 67–72. [Google Scholar]
  3. Yan, Z.; Xiao, Y.; Cheng, L.; He, R.; Ruan, X.; Zhou, X.; Li, M.; Bin, R. Exploring AIS data for intelligent maritime routes extraction. Appl. Ocean. Res. 2020, 101, 102271. [Google Scholar] [CrossRef]
  4. United States Coast Guard AIS Encoding Guide. Available online: https://www.navcen.uscg.gov/pdf/AIS/AISGuide.pdf (accessed on 16 February 2022).
  5. Jakovlev, S.; Daranda, A.; Voznak, M.; Lektauers, A.; Eglynas, T.; Jusis, M. Analysis of the Possibility to Detect Fake Vessels in the Automatic Identification System. In Proceedings of the 2020 61st International Scientific Conference on Information Technology and Management Science of Riga Technical University (ITMS), Riga, Latvia, 15–16 October 2020; pp. 1–5. [Google Scholar]
  6. Struck, M.C.; Stoppe, J. A Backwards Compatible Approach to Authenticate Automatic Identification System Messages. In Proceedings of the 2021 IEEE International Conference on Cyber Security and Resilience (CSR), Rhodes, Greece, 26–28 July 2021; pp. 524–529. [Google Scholar]
  7. Wimpenny, G.; Safar, J.; Grant, A.; Bransby, M.; Ward, N. Public key authentication for AIS and the VHF data exchange system (VDES). In Proceedings of the 31st International Technical Meeting of the Satellite Division of The Institute of Navigation (ION GNSS+ 2018), Miami, FL, USA, 24–28 September 2018; pp. 1841–1851. [Google Scholar]
  8. Alincourt, E.; Ray, C.; Ricordel, P.M.; Dare-Emzivat, D.; Boudraa, A. Methodology for AIS signature identification through magnitude and temporal characterization. In Proceedings of the OCEANS 2016-Shanghai, Shanghai, China, 10–13 April 2016; pp. 1–6. [Google Scholar]
  9. Balduzzi, M.; Pasta, A.; Wilhoit, K. A security evaluation of AIS automated identification system. In Proceedings of the 30th Annual Computer Security Applications Conference, New Orleans, LA, USA, 8–12 December 2014; pp. 436–445. [Google Scholar]
  10. Li, F.; Chen, C.H.; Xu, G.; Chang, D.; Khoo, L.P. Causal factors and symptoms of task-related human fatigue in vessel traffic service: A task-driven approach. J. Navig. 2020, 73, 1340–1357. [Google Scholar] [CrossRef]
  11. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  12. Prasad, D.K.; Rajan, D.; Rachmawati, L.; Rajabally, E.; Quek, C. Video processing from electro-optical sensors for object detection and tracking in a maritime environment: A survey. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1993–2016. [Google Scholar] [CrossRef] [Green Version]
  13. Shao, Z.; Wu, W.; Wang, Z.; Du, W.; Li, C. Seaships: A large-scale precisely annotated dataset for ship detection. IEEE Trans. Multimed. 2018, 20, 2593–2604. [Google Scholar] [CrossRef]
  14. Chen, X.; Qi, L.; Yang, Y.; Luo, Q.; Postolache, O.; Tang, J.; Wu, H. Video-based detection infrastructure enhancement for automated ship recognition and behavior analysis. J. Adv. Transp. 2020, 2020, 7194342. [Google Scholar] [CrossRef] [Green Version]
  15. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  16. Qiao, S.; Chen, L.C.; Yuille, A. Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10213–10224. [Google Scholar]
  17. Bolya, D.; Zhou, C.; Xiao, F.; Lee, Y.J. Yolact: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9157–9166. [Google Scholar]
  18. Lee, Y.; Park, J. Centermask: Real-time anchor-free instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13906–13915. [Google Scholar]
  19. Zhao, H.; Zhang, W.; Sun, H.; Xue, B. Embedded Deep Learning for Ship Detection and Recognition. Future Internet 2019, 11, 53. [Google Scholar] [CrossRef] [Green Version]
  20. Ghahremani, A.; Kong, Y.; Bondarev, E. Multi-class detection and orientation recognition of vessels in maritime surveillance. Electron. Imaging 2019, 2019, 266-1–266-5. [Google Scholar] [CrossRef]
  21. Nita, C.; Vandewal, M. CNN-based object detection and segmentation for maritime domain awareness. In Artificial Intelligence and Machine Learning in Defense Applications II; International Society for Optics and Photonics: Bellingham, WA, USA, 2020; Volume 11543, p. 1154306. [Google Scholar]
  22. Han, K.M.; DeSouza, G.N. Geolocation of multiple targets from airborne video without terrain data. J. Intell. Robot. Syst. 2011, 62, 159–183. [Google Scholar] [CrossRef]
  23. Cai, Y.; Ding, Y.; Xiu, J.; Zhang, H.; Qiao, C.; Li, Q. Distortion measurement and geolocation error correction for high altitude oblique imaging using airborne cameras. J. Appl. Remote Sens. 2020, 14, 014510. [Google Scholar] [CrossRef]
  24. El Habchi, A.; Moumen, Y.; Zerrouk, I.; Khiati, W.; Berrich, J.; Bouchentouf, T. CGA: A New Approach to Estimate the Geolocation of a Ground Target from Drone Aerial Imagery. In Proceedings of the 2020 Fourth International Conference On Intelligent Computing in Data Sciences (ICDS), Fez, Morocco, 21–23 October 2020; pp. 1–4. [Google Scholar]
  25. Gao, F.; Deng, F.; Li, L.; Zhang, L.; Zhu, J.; Yu, C. MGG: Monocular Global Geolocation for Outdoor Long-Range Targets. IEEE Trans. Image Process. 2021, 30, 6349–6363. [Google Scholar] [CrossRef] [PubMed]
  26. Naus, K.; W, M.; Szymak, P.; Gucma, L.; Gucma, M. Assessment of ship position estimation accuracy based on radar navigation mark echoes identified in an Electronic Navigational Chart. Measurement 2020, 169, 108630. [Google Scholar] [CrossRef]
  27. Liu, R.W.; Yuan, W.; Chen, X.; Lu, Y. An enhanced CNN-enabled learning method for promoting ship detection in maritime surveillance system. Ocean. Eng. 2021, 235, 109435. [Google Scholar] [CrossRef]
  28. Svanberg, M.; Santén, V.; Hörteborn, A.; Holm, H.; Finnsgård, C. AIS in maritime research. Mar. Policy 2019, 106, 103520. [Google Scholar] [CrossRef]
  29. Helgesen, Ø.K.; Brekke, E.F.; Stahl, A.; Engelhardtsen, Ø. Low Altitude Georeferencing for Imaging Sensors in Maritime Tracking. IFAC-Pap. 2020, 53, 14476–14481. [Google Scholar] [CrossRef]
  30. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  31. Wada, K. labelme: Image Polygonal Annotation with Python. 2016. Available online: https://github.com/wkentaro/labelme (accessed on 16 February 2022).
  32. Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the Protection of Natural Persons with Regard to the Processing of Personal Data and on the Free Movement of Such Data, and Repealing Directive 95/46/EC (General Data Protection Regulation) (Text with EEA Relevance). 2016. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32016R0679 (accessed on 16 February 2022).
  33. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  35. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  38. Guo, Y.; Chen, F.; Cheng, Q.; Wu, J.; Wang, B.; Wu, Y.; Zhao, W. Fully Convolutional One-Stage Circular Object Detector on Medical Images. In Proceedings of the 2020 4th International Conference on Advances in Image Processing, Chengdu, China, 13–15 November 2020; pp. 21–26. [Google Scholar]
  39. Lee, Y.; Hwang, J.w.; Lee, S.; Bae, Y.; Park, J. An energy and gpu-computation efficient backbone network for real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  40. Solano-Carrillo, E.; Carrillo-Perez, B.; Flenker, T.; Steiniger, Y.; Stoppe, J. Detection and Geovisualization of Abnormal Vessel Behavior from Video. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021. [Google Scholar]
  41. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  42. Wu, Y.; Kirillov, A.; Massa, F.; Lo, W.Y.; Girshick, R. Detectron2. 2019. Available online: https://github.com/facebookresearch/detectron2 (accessed on 16 February 2022).
  43. Pawlowski, E. Experimental study of a positioning accuracy with GPS receiver. In Proceedings of the 12th Conference on Selected Problems of Electrical Engineering and Electronics, WZEZ, Kielce, Poland, 17–19 September 2015. [Google Scholar] [CrossRef]
Figure 1. View of each camera used for ShipSG data collection, showing the river, port basin and lock entrance. (a) View of first camera. (b) View of second camera.
Figure 1. View of each camera used for ShipSG data collection, showing the river, port basin and lock entrance. (a) View of first camera. (b) View of second camera.
Sensors 22 02713 g001
Figure 2. Examples extracted from the dataset that show the seven ship classes. Each class contains a variety of sizes and orientations of the ships. (a) Cargo , (b) Tug, (c) Special 1, (d) Tanker, (e) Law Enforcement, (f) Passenger/Pleasure, (g) Special 2.
Figure 2. Examples extracted from the dataset that show the seven ship classes. Each class contains a variety of sizes and orientations of the ships. (a) Cargo , (b) Tug, (c) Special 1, (d) Tanker, (e) Law Enforcement, (f) Passenger/Pleasure, (g) Special 2.
Sensors 22 02713 g002
Figure 3. Visualisation of ShipSG dataset samples with annotated ship masks and classes, and one ship position per image.
Figure 3. Visualisation of ShipSG dataset samples with annotated ship masks and classes, and one ship position per image.
Sensors 22 02713 g003
Figure 4. Reference pairs to create the homographies. The coloured dots to the left show the latitudes and longitudes obtained from the AIS messages on a map. To the right, the counterpart pixel coordinates are displayed, annotated by hand. The black icons show the location of the cameras. (a) First camera pairs for homography creation, with blue dots. (b) Second camera pairs for homography creation, with red dots.
Figure 4. Reference pairs to create the homographies. The coloured dots to the left show the latitudes and longitudes obtained from the AIS messages on a map. To the right, the counterpart pixel coordinates are displayed, annotated by hand. The black icons show the location of the cameras. (a) First camera pairs for homography creation, with blue dots. (b) Second camera pairs for homography creation, with red dots.
Sensors 22 02713 g004
Figure 5. Example of segmented ship mask with calculated pixel to be georeferenced (in red, enlarged for visualisation).
Figure 5. Example of segmented ship mask with calculated pixel to be georeferenced (in red, enlarged for visualisation).
Sensors 22 02713 g005
Figure 6. Annotated masks on existing datasets for the study of the generalisation of our models. (a) Annotated examples of the SMD [12]. (b) Annotated examples of Seaships7000 [13]. (c) Annotated examples of the dataset by Chen et al. [14].
Figure 6. Annotated masks on existing datasets for the study of the generalisation of our models. (a) Annotated examples of the SMD [12]. (b) Annotated examples of Seaships7000 [13]. (c) Annotated examples of the dataset by Chen et al. [14].
Sensors 22 02713 g006
Figure 7. Qualitative ship georeferencing results. (a) Port basin. (b) River. Green dots show the positions given by AIS. Orange dots show the georeferenced positions using our method). Gray lines join the actual and georeferenced positions.
Figure 7. Qualitative ship georeferencing results. (a) Port basin. (b) River. Green dots show the positions given by AIS. Orange dots show the georeferenced positions using our method). Gray lines join the actual and georeferenced positions.
Sensors 22 02713 g007
Figure 8. Georeferencing distance error per ship length. GDEs and their uncertainties fall within the bounds of the ship length.
Figure 8. Georeferencing distance error per ship length. GDEs and their uncertainties fall within the bounds of the ship length.
Sensors 22 02713 g008
Table 1. Number of masks annotated per class for each area scale.
Table 1. Number of masks annotated per class for each area scale.
ClassSmallMediumLargeTotal%
Cargo98902300130011.18
Law Enforcement1013536111374832.24
Passenger/Pleasure48485936265.38
Special 16442751110028.62
Special 2265231253263022.62
Tanker277753382141212.15
Tug2494701889077.80
All classes11028885163811625100
Table 2. Configurations during training for each method evaluated.
Table 2. Configurations during training for each method evaluated.
MethodInput Size (Pixel)BackboneBatch SizeIterationsNumber of Epochs
Mask R-CNN [15]1333 × 800ResNeXt-101215,40011
DetectoRS [16]1333 × 800ResNet-50215,40011
YOLACT550 [17]550 × 550ResNet-508648018
YOLACT700 [17]700 × 700ResNet-1018576016
Centermask-LiteV19 [18]800 × 600Vovnet-198594917
Centermask-LiteV39 [18]800 × 600Vovnet-398594917
Table 3. Resulting instance segmentation APs and inference speed per method evaluated. The two first rows are robust methods and the rest are the real-time methods.
Table 3. Resulting instance segmentation APs and inference speed per method evaluated. The two first rows are robust methods and the rest are the real-time methods.
MethodAPcamAPmAP50mAP75mAPSmAPMmAPLFPS
Mask R-CNN [15]0.7720.7330.9610.9140.5030.7520.7728.50
DetectoRS [16]0.7800.7470.9820.9240.5560.7570.7926.62
YOLACT550 [17]0.5710.5270.8860.6090.0860.5150.70936.28
YOLACT700 [17]0.6220.5820.9110.7000.1400.5820.75127.75
Centermask-LiteV19 [18]0.7320.6350.8400.7800.4550.6400.65740.98
Centermask-LiteV39 [18]0.7400.6440.8390.7870.4610.6480.66135.25
Table 4. Generalisation mask AP results per model, the first two rows are robust methods and the rest are real-time capable methods. All the classes have been contemplated as a single class (class-agnostic).
Table 4. Generalisation mask AP results per model, the first two rows are robust methods and the rest are real-time capable methods. All the classes have been contemplated as a single class (class-agnostic).
MethodAPAP50
Mask R-CNN [15]0.4410.749
DetectoRS [16]0.4860.830
YOLACT550 [17]0.3400.615
YOLACT700 [17]0.3360.613
Centermask-LiteV19 [18]0.3480.636
Centermask-LiteV39 [18]0.3870.710
Table 5. Quantitative ship georeferencing results. Δ φ and Δ λ stand for absolute latitude and longitude error, respectively. GDE stands for georeferencing distance error. Std stands for standard deviation. DRSME stands for distance root-mean-square error.
Table 5. Quantitative ship georeferencing results. Δ φ and Δ λ stand for absolute latitude and longitude error, respectively. GDE stands for georeferencing distance error. Std stands for standard deviation. DRSME stands for distance root-mean-square error.
LocationMean  Δ φ [m]Mean  Δ λ [m]Mean GDE [m]Std GDE [m]DRSME [m]
Port Basin (range < 400 m)1612221027
River (range > 400 m)4227532461
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Carrillo-Perez, B.; Barnes, S.; Stephan, M. Ship Segmentation and Georeferencing from Static Oblique View Images. Sensors 2022, 22, 2713. https://doi.org/10.3390/s22072713

AMA Style

Carrillo-Perez B, Barnes S, Stephan M. Ship Segmentation and Georeferencing from Static Oblique View Images. Sensors. 2022; 22(7):2713. https://doi.org/10.3390/s22072713

Chicago/Turabian Style

Carrillo-Perez, Borja, Sarah Barnes, and Maurice Stephan. 2022. "Ship Segmentation and Georeferencing from Static Oblique View Images" Sensors 22, no. 7: 2713. https://doi.org/10.3390/s22072713

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop