Next Article in Journal
Creative Environments for Sustainability in Organizations Promoting the Education of Vulnerable Groups
Previous Article in Journal
Assessment of Factors Influencing the Development of Inland Navigation in Poland
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Identification of Sunken Oil in Homogeneous Information Perturbed Environment through Fusion Image Enhancement with Convolutional Neural Network

1
School of Mechanical & Automotive Engineering, Qingdao University of Technology, Qingdao 266520, China
2
Guangdong Provincial Key Laboratory of Deep Earth Sciences and Geothermal Energy Exploitation and Utilization, Institute of Deep Earth Sciences and Green Energy, College of Civil and Transportation Engineering, Shenzhen University, Shenzhen 518060, China
3
School of Energy and Mining Engineering, China University of Mining & Technology, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
Sustainability 2024, 16(15), 6665; https://doi.org/10.3390/su16156665
Submission received: 11 July 2024 / Revised: 29 July 2024 / Accepted: 1 August 2024 / Published: 4 August 2024
(This article belongs to the Section Waste and Recycling)

Abstract

:
With the development of the marine oil industry, leakage accidents are one of the most serious problems threatening maritime and national security. The spilt crude oil can float and sink in the water column, posing a serious long-term threat to the marine environment. High-frequency sonar detection is currently the most efficient method for identifying sunken oil. However, due to the complicated environment of the deep seabed and the interference of the sunken oil signals with homogeneous information, sonar detection data are usually difficult to interpret, resulting in low efficiency and a high failure rate. Previous works have focused on features designed by experts according to the detection environments and the identification of sunken oil targets regardless of the feature extraction step. To automatically identify sunken oil targets without a prior knowledge of the complex seabed conditions during the image acquisition process for sonar detection, a systematic framework is contrived for identifying sunken oil targets that combines image enhancement with a convolutional neural network (CNN) classifier for the final decision on sunken oil targets examined in this work. Case studies are conducted using datasets obtained from a sunken oil release experiment in an outdoor water basin. The experimental results show that (i) the method can effectively distinguish between the sunken oil target, the background, and the interference target; (ii) it achieved an identification accuracy of 83.33%, outperforming feature-based recognition systems, including SVM; and (iii) it provides important information about sunken oil such as the location of the leak, which is useful for decision-making in emergency response to oil spills at sea. This line of research offers a more robust and, above all, more objective option for the difficult task of automatically identifying sunken oils under complex seabed conditions.

1. Introduction

With the continuous expansion of offshore oil production and utilization, the number of oil spills around the world has gradually increased. For example, between 1973 and 2020, there were 156 oil spills involving vessels with a capacity of more than 10 tons in coastal areas of China [1]. In 2010, an oil spill occurred in the US Gulf of Mexico, which caused great concern in the international community [2]. Marine oil spill accidents not only lead to the waste of marine oil resources but also cause environmental pollution of varying degrees, which poses an immediate threat to marine safety [3]. In the event of a large oil spill, the spilled crude oil and heavy oil may float in the water column and sink due to factors such as oil density, water turbulence caused by waves or currents, and sediment floating in the water [4]. Due to changes in temperature, hydraulics, and the marine environment, the sunken oil may be dispersed by ocean wave currents and float to the sea surface or to land seasonally, resulting in secondary pollution. As a result, sunken oil can impact the marine environment for longer periods of time and over greater distances than floating oil. Therefore, it is critical to detect and identify sunken oil at the first attempt after an oil spill and take corrective emergency measures to completely remove it.
Recent research has explored various techniques for detecting sunken oil. Visible light and SAR satellite remote sensing effectively detect floating oil, but their effectiveness decreases with depth due to wavelength attenuation. This limitation has spurred research into alternative detection techniques. Diver observations, for example, offer direct and accurate results but are restricted by depth, visibility, and human endurance [5]. Laser fluorosensors can detect oil on or near the water surface, but their effectiveness decreases with depth, limiting their use for sunken oil detection [6]. In contrast, bottom trawls are effective for detecting oil on the seafloor but potentially damaging to benthic ecosystems [7]. Comparative studies have evaluated these methods based on efficiency, time consumption, cost, and operational limitations. Experimental research has demonstrated that acoustic technology, particularly high-frequency sonar, outperforms other methods for sunken oil detection [8]. In 2005, an oil spill occurred in the US Gulf of Mexico, and as the oil spill area gradually expanded, side-scan sonar technology became the only way to detect sunken oil [9]. In 2015, an accident occurred in the Mississippi River where two oil tankers collided, causing tons of crude oil to spill into the river downstream. The side-scan sonar successfully located two sunken oil targets [10]. Given its demonstrated effectiveness and ongoing technological improvements, acoustic technology, specifically high-frequency sonar, positions itself as the most promising solution for addressing the challenges of sunken oil detection.
Despite these advances, challenges remain in distinguishing sunken oil from other seafloor features and accurately quantifying oil volumes. Current research focuses on integrating multiple sensing technologies and developing advanced data processing algorithms to address these issues. Recent advancements in sonar technology have further improved detection accuracy. For instance, Snellen et al. [11] developed a multi-beam sonar system that achieved seabed sediment classification in controlled experiments. Similarly, Tang et al. [12] combined high-frequency sonar with neural networks, enhancing classification accuracy to 88% in field trials. The above methods are effective when it comes to detecting sunken oil in shallow waters. As the depth of the water increases, the signal for detecting sunken oil is not only significantly weakened, but it also becomes mixed up with signals from other objects floating in the water and on the seabed. Therefore, interpreting and identifying the sunken oil heavily relies on the expert’s prior knowledge. In this context, developing an automated approach to identify the sunken oil under complex deep seabed conditions is feasible and critical to supporting emergency response to offshore oil spills.
In recent years, great progress has been made in automatically identifying underwater targets based on sonar detection images, thanks to advances in the machine learning and deep learning communities. For example, Bian et al. used the pixel value importance functions to initially filter the targets; they then used a CNN as a secondary classification method to achieve underwater target identification [13]. Wang et al. proposed a novel deep learning model using an adaptive weighting convolutional neural network (AW-CNN) to classify underwater sonar detection images [14]. Williams et al. used convolutional neural networks to perform underwater target classification in Synthetic Aperture Sonar (SAS) images [15]. However, the following challenges exist when using computer vision methods to detect and identify sunken oil: (i) the image conveys more complicated information, such as background with sand and other sediments, and (ii) the sunken oil targets vary in size and shape. Therefore, more powerful approaches need to be developed to accurately identify the information about sunken oil targets.
Currently, there are few studies on the application of deep learning methods and CNN algorithms to identify sunken oil. This paper presents a novel sunken oil identification scheme based on CNN. This multidisciplinary study integrates underwater high-frequency sonar detection technology, image processing technology, and CNN algorithms. In addition, the present study improves the technical capabilities of marine oil spill detection and identification methods and provides technical support for marine safety.
The remaining parts of this paper are organized as follows: Section 2 explains the proposed framework for sunken oil identification including the image preprocessing methods for sonar detection and the CNN architecture. Section 3 presents the details of sunken oil release experiments and case studies on experiment data to evaluate the effectiveness of the framework. The conclusions are summarized in Section 4.

2. Proposed Method for Sunken Oil Identification

Due to the background noise and other interfering targets, the sonar detection image has a low signal-to-noise ratio and low resolution. Therefore, this paper proposes a systematic framework for sunken oil identification based on CNN; the proposed method is divided into two stages to accurately identify the sunken oil target. The first step is to preprocess the original sonar detection image. The goal of this phase is to minimize background noise in the image and segment the suspected sunken oil target areas (including actual sunken oil targets and other nuisance targets). In the second stage, a pre-trained CNN classification model is used to classify and identify the suspected sunken oil target areas, exclude background and interference targets, and determine the actual sunken oil target. The detailed algorithm implementation process is shown in Figure 1.

2.1. Stage i: Homogeneous Interference Signal Attenuation and Target Signal Amplification

The preprocessing of the attenuation of the homogeneous interference signal and target signal amplification includes three processes: homogeneous image filtering, target image amplification, and threshold segmentation, which are described in detail below.
Gaussian filtering is used for homogeneous interference signal suppression [16]. Assuming the size of the filter kernel is n × n, then the two-dimensional Gaussian filter G ( i ,   j ) is defined by (1).
G i , j = 1 2 π σ 2 e i 2 + j 2 / 2 σ 2 where   i     [ 0 ,   n ] ,   j     [ 0 ,   n ]
where σ is the standard deviation of the normal distribution. The output of Gaussian filtering is the Gaussian-weighted average of the gray values of all pixels in the filter kernel. The calculation for Gaussian filtering is shown in (2).
F i , j = 1 i = 1 n j = 1 n G i , j i = 1 n j = 1 n G i , j g ( i , j )
where G ( i ,   j ) is the gray value of pixels at the location described as a coordinate (i, j) in the filter kernel, g ( i , j ) is the gray value of the input image pixel at coordinates (i, j) within the filter window, and F i , j is the gray value of the pixel in the middle of the filter kernel.
In this study, the Gaussian filter with a filter kernel of 35 × 35 is used to filter sonar detection images. Figure 2b shows the image filtering results obtained when the Gaussian filter with a 35 × 35 filter kernel was applied to the original image. It is clear that the background noise of the filtered image has been largely eliminated.
It is clear from Figure 2b that the filtered image reduces the local contrast of the image and, unfortunately, by suppressing the background noise, the signal in the submerged oil target area is also weakened. Therefore, it is necessary to improve the image. This article applies the histogram equalization algorithm. The algorithm converts the grayscale histogram of the input image from a specific grayscale interval in the comparison set to a uniform distribution throughout the grayscale range, thereby improving the image contrast, improving the image dynamic range, and making the image information more prominent [17]. The effect of improving the sonar detection image is shown in Figure 3. Compared to Figure 2b, the sunken oil area in Figure 3 showed greater contrast to the background image.
After the improved processing, some useless information was incorrectly improved. Therefore, it is also necessary to perform threshold segmentation on the image to segment the sunken oil target area from the background. The sunken oil appears primarily as dark spots compared to the background (see Figure 2b). This feature suggests the use of a threshold segmentation method [18]. The current main approach to threshold segmentation can be categorized into fixed threshold segmentation, adaptive threshold segmentation (adaptive mean threshold segmentation and adaptive Gaussian threshold segmentation), and the Otsu method [19].
The grayscale histogram of the sonar detection images showed a Gaussian distribution; therefore, a fixed threshold segmentation approach was used, and the segmentation threshold was set to the tertile of the grayscale distribution. Next, the image was eroded to obtain a more accurate segmentation result. The post-erosion binary segmentation image is shown in Figure 4.
After threshold segmentation, the Canny operator [20] is used to detect the edge of dark areas in Figure 4 (as shown in Figure 5). Then, the outer rectangle is drawn from dark areas. These areas are segmented from the original image and treated as suspected sunken oil targets because they may contain both the sunken oil targets and the background. The segmented areas are then identified using the CNN model (proper scaling is applied to adjust the CNN input dimensions) to exclude the background area and make final decisions. The extracted target area is shown in Figure 6.

2.2. Stage ii: CNN-Based Sunken Oil Detection

2.2.1. Brief Description of Convolutional Neural Network

Recently, deep learning methods have been successfully applied to increasingly complex practical engineering problems, including target detection [21], fault diagnosis [22], and condition monitoring [23]. In addition, convolutional neural network algorithms provide an end-to-end learning model and are an “excellent representative” of deep learning methods, which are widely used in tasks such as image classification [24] and object detection [25]. Through multiple convolution layers and pooling operations, CNNs automatically learn features from input images; they then input these features into the fully connected neural network to achieve image identification and classification. The structure of a convolutional neural network usually consists of convolution layers, pool layers, fully connected layers, and output layers.
The convolution layer convolves the input data using convolution filters. According to [26], the convolution operation for the n-th feature map in the i-th convolutional layer is shown in Equation (3).
H n i = f m = 1 M H m i 1 W n , m i + b n i
where W n , m i is the weight kernel connecting the m-th feature map in layer (i − 1) to the n-th feature map in layer i, H m i 1 is the m-th feature map from the previous layer (i − 1), and H n i is the n-th feature map in the i-th layer. The two-dimensional convolution operation of the convolution filter of the i-th layer and the feature map of the last layer is represented by . The bias vector b n i is added to the result of the convolution operation. This bias term allows the network to shift the activation function, providing flexibility in fitting the data. It acts as a trainable offset, enabling the model to represent patterns that do not necessarily pass through the origin. The nonlinear activation function f introduces nonlinearity into the network, allowing it to learn complex patterns. Common choices for f include the Rectified Linear Unit (ReLU), which is defined as f ( x ) = m a x ( 0 ,   x ) , or the sigmoid function, f x = 1 / 1 + e x . As the gradient of the ReLU is piecewise-constant, in contrast with the sigmoid function, the ReLU can effectively avoid a vanishing gradient when the activation tends to be positively infinite. Therefore, this study uses the ReLU as the activation function; the definition of the ReLU function is shown in Equation (4).
relu x = 0 ,     x 0 x ,     x > 0  
where x is the activation value.
The pooling layer, also called the subsampling layer, aims to achieve spatial invariance by reducing the size of the feature maps. Typical pooling operations are average pooling and maximum pooling. Average pooling means finding the average values in each sub-range of the input data, and maximum pooling means collecting the maximum value in each sub-range of the input data. The pooling operation adopted by all pooling layers in this article is maximum pooling.
The fully connected layer is similar to the classic artificial neural network, where the neurons in each hidden layer are connected to each neuron in the previous layer. The last layer of the fully connected layer is also called the classifier layer, which calculates a class-wise posterior probability of the input image. The output is calculated using the softmax regression function [27] shown in (5).
p y = i x ; θ = e θ i x j = 1 K e θ j x
where K is the number of predicted classes and θ i x is the i-th activation value of the classifier layer.

2.2.2. CNN Method for Sunken Oil Identification

The structure of the CNN model used in this work is shown in the right panel of Figure 1 and includes a convolutional layer, a pooling layer, and a fully connected layer.
As shown in Table 1, in the convolution layer, the size of the convolution filter is set to 5 × 5 and the stride is set to 1 pixel; the number of filters is 5. The output includes five feature maps with a size of 71 × 71. In the pooling layer, maximum pooling is performed over a 2 × 2 window with a step size of 2. The output is 5 feature maps with a size of 35 × 35. The fully connected layer consists of a hidden layer with 30 neurons. The final output of the model is a one-dimensional vector with three elements, and the three element values represent the confidence that the input image is judged as a sunken oil target, interference target, and background. The CNN uses the softmax function for classification and the ReLU as the activation function. We employed categorical cross-entropy as our loss function due to its effectiveness in multi-class classification problems, which aligns with our task of distinguishing between sunken oil targets and various types of interference. During training, a batch size ranging from 6 to 24 is used and the learning rate for Adam optimization [28] is chosen between 10−3 and 10−4. The batch size and learning rate are selected using cross-validation. An early stop with a patience of 6 is implemented.

2.3. Comparison Method

The Support Vector Machine (SVM) is used as a comparison method in this work. SVM is considered a widely used shallow method that has been proven to be efficient and useful for classifying and detecting oil leakage in SAR images [29,30]. The goal of SVM is to find an optimal hyperplane that can separate two sets of data points while maximizing the distance between the hyperplane and the data points [31].
The radial basis function (RBF) kernel is adopted for SVM training in this work. And, the hyperparameters are searched in the training set using the cross-validation method.
This article uses two evaluation methods, precision rate (P) and recall rate (R), to evaluate the performance of the proposed method. The precision rate (P) is defined as the proportion of correctly predicted samples in the prediction results for each category; the recall rate (R) is defined as the proportion of correctly predicted samples from the original samples for each category. The definition of precision rate and recall rate is shown in (6) and (7).
P = T P T P + F P × 100
R = T P T P + F N × 100
where T P is the number of true positives, F N is the number of false negatives, and F P is the number of false positives. Intersection over union (IoU) is used to decide whether identification regions are true or false positives by measuring the overlap area with ground truth [32]. The definition of IoU is shown in (8), where P and G T are the area of detection and ground truth region, respectively.
I o U = P G T P G T  
where P G T denotes the intersection of the detection and ground truth region and P G T their union. In this paper, identifications are considered true positives ( T P s) and assigned to the corresponding ground truth annotations if the IoU exceeds 0.3. Ground truth objects not detected are false negatives ( F N s). Multiple identifications of the same target or identifications with no matching ground truth are considered false positives ( F P s).
Besides the precision rate and recall rate, the F-score is applied to further summarize the performance of the proposed method. A general definition of F-score [33] is given in (9).
F β = 1 + β 2 P R β 2 P + R × 100
where P represents the precision rate, R represents the recall rate, and β is a parameter that determines the weight of precision in the combined score. When β = 1 , precision and recall are weighted equally, resulting in the F 1 score.

3. Experiments and Discussions

3.1. Experimental Setup

In order to obtain the sonar detection images for CNN training, a series of experiments were first conducted in the outdoor swimming pool of CNOOC Energy Development Co., Ltd., Department of Safety and Environmental Protection, Tianjin, China. The schematic diagram of the experimental setup is shown in Figure 7.
The structure consists of the outdoor pool, the slider, and the high-frequency sonar. The outdoor pool is 2 m deep, 5 m long, and 1.8 m wide. The slider attached to the edge of the pool moves in both horizontal and vertical directions. A BlueView M900-2250 imaging sonar (technical details listed in Table 2) is attached to the slider to scan the bottom of the tank at various angles and depths during the experiment. The basin is filled with 1.8 m of deep seawater, which had a conductivity of 47.1 mS/cm and a density of 1023 kg/m³, respectively. All experiments are carried out at a water temperature of Twater = 16 °C. The bottom of the pool is covered with sea sand. Crude oil (physical and chemical properties are listed in Table 3) is mixed with kaolin to simulate sunken oil sediment targets and placed at the bottom of the basin along with other interference targets to simulate the complex seafloor environment. The settings for the experimental conditions at the bottom of the pool are listed in Table 4. An example of the obtained sonar detection image is shown in Figure 8.

3.2. Data Preparation

A total of 58 sonar detection images were collected as part of the experiment. The location of sunken oil targets and interference targets in sonar detection images is commented on by experts and considered ground truth. Image reasoning methods (e.g., changing the contrast and brightness of the image and adding noise) are then applied to generate further image data. Finally, 104 images are obtained. Eighty-two sonar detection images are selected to pre-train the CNN model, and each ground truth region is cropped into small sample images according to a 75 × 75 cropping window. The crop window moves from the top left to the bottom right of the image in 15-pixel increments. After each process, 50 to 60 sample images per original image are obtained. A total of 1963 samples of submerged oil paintings, 1628 samples of interference target images, and 1447 samples of background images are obtained. Illustrative examples are shown in Figure 9.

3.3. CNN Pre-Training and Results

Before being implemented into the detection framework, the proposed CNN is pre-trained on the dataset to learn the features of sunken oil. In Section 3.2, a total of 5038 sample images were cropped, from which 3903 sample images were randomly selected for training and testing the CNN model. Of these, 3399 sample images are used for CNN model training and 504 sample images are used for model testing. The CNN model was implemented using Python 3.6 and TensorFlow 2.6 on a Windows 10 operating system. All experiments were conducted on a high-performance workstation equipped with an NVIDIA RTX 2080 Ti GPU (with CUDA 11) and an Intel Xeon Silver 4210R CPU. Figure 10 shows the loss value curve of the training set and the validation set in the first 14 training epochs. The loss was calculated using categorical cross-entropy, which is appropriate for our multi-class classification task. As the number of training epochs increases, the categorical cross-entropy loss values for both the training and validation sets gradually approach zero. Figure 11 is a comparison of the gray values of the 10 convolution kernels before and after training, showing that the convolution kernels maintain certain rules. Most convolution kernels respond significantly to the boundary texture of submerged oil targets, which is consistent with the observations in the previous experiment.
Table 5 shows the confusion matrix of the identification results predicted by the CNN. It is clear that the samples of the sunken oil target that were misidentified were all incorrectly detected as interference targets. The main reason for this misidentification is that the characteristics of some interference targets are similar to those of the sunken oil targets (as shown in Figure 9a,b). The overall detection accuracy of the proposed model (the proportion of correctly identified samples out of the total number of test samples) is 97.62%, indicating that the proposed CNN model is accurate and robust in identifying the sunken oil targets.
To further explain the identification results, Table 6 shows the precision rates, recall rates, and F1 scores of the three categories. It is obvious that the precision rates of the three categories are all above 94%; However, the precision and recall rates of the sunken oil target are the highest (98.99% and 98.01%, respectively). This result indicates that the sunken oil targets have significantly different characteristics compared to the other two target types. The corresponding histogram is shown in Figure 12.
Please note that the interference target classification result shown here is only to indicate the CNN classification ability. In the further process of identifying sunken oil, the identified disturbing targets and their backgrounds are eliminated.

3.4. Sunken Oil Identification

With the trained CNN model, 22 test images are input into the identification framework to detect the sunken oil targets. Examples of sunken oil detection and identification are shown in Figure 13b. The sunken oil targets in Figure 13a are all correctly identified. Their position is clearly marked by red bounding boxes. Meanwhile, two false positives were detected and marked with yellow bounding boxes. The number of identified targets and false alarms in each image is counted and displayed in the upper-right corner of each image. As the results show, the proposed framework can identify sunken oil targets and provide important information such as location to support further emergency decisions.
The performance of the proposed method and the comparison SVM method are shown in Table 7. The SVM model is trained using gray values and the Grey Level Co-occurrence Matrix (GLCM) [34] features extracted from the same training image set used by CNN. Please note that precision and recall are two important aspects when evaluating performance. In this case, it is important to place more emphasis on the recall because undetected sunken oil can cause greater damage than a false alarm. Based on the results, the identification framework successfully identified 83.33% of the sunken oil targets, outperforming the other two SVM models (80.56% and 75%, respectively). The results in Table 7 are consistent with the target and only a few false positives (76 to 16) were detected. In addition, the identification model has a stable performance in both training and testing. The result showed that the identification framework can identify sunken oil targets with reliable accuracy.

4. Discussions

4.1. Automatic Identification for Sunken Oil Detection Applicability Using CNN

The automatic identification of sunken oil in this study is based on a CNN classifier trained on enhanced sonar images. The CNN model leverages deep learning techniques to extract and learn features from the input images, allowing it to distinguish between different classes, such as sunken oil, background, and interference targets. Image enhancement techniques were applied to the raw sonar images to improve their quality and highlight relevant features. This preprocessing step is crucial for improving the CNN model’s performance, as it enhances the visibility of sunken oil and other targets, making them easier to detect. The CNN model learns hierarchical features from the enhanced images through multiple layers of convolutional filters. These filters automatically extract and represent relevant patterns and characteristics, such as edges, textures, and shapes, which differentiate sunken oil from other targets. After feature extraction, the CNN model classifies the input images into predefined categories based on the learned features. The classification output includes probability scores for each class, allowing the model to make informed decisions about the presence of sunken oil.
The CNN model achieved an overall F1 score of 81.08%, significantly higher than the SVM model, which had identification accuracies of 78.38% for sunken oil targets. The precision rates for identifying background, interference targets, and sunken oil targets using CNN exceeded 94%, with sunken oil targets having the highest precision (98.99%) and recall (98.01%). This high accuracy level ensures reliable sunken oil detection, which is critical for environmental monitoring and emergency response. The experiment demonstrated the CNN model’s robustness in complex seabed conditions, effectively distinguishing between sunken oil, background, and interference targets. This adaptability to complex conditions enhances the model’s performance and reliability, making it a valuable tool in marine environmental monitoring and emergency response.
However, the CNN model requires substantial labeled training data to achieve high accuracy. This study employed data augmentation techniques to enhance the training dataset, generating 5038 sample images. The need for extensive training data can be a limitation in real-world scenarios where obtaining labeled data is challenging.
The developed approach has significant potential for practical applications in automatically detecting sunken oil in marine environments. The CNN model’s integration into sonar imaging systems deployed on underwater vehicles, such as Autonomous Underwater Vehicles (AUVs) or Remotely Operated Vehicles (ROVs), can enable the real-time monitoring of marine environments, providing the continuous and automated detection of sunken oil. This potential for practical application underscores the value and relevance in the field of marine environmental monitoring and emergency response.

4.2. Experimental Settings

The experimental settings employed in this study played a crucial role in validating the effectiveness of the proposed method. The controlled setting allowed for the precise manipulation of variables, including the placement of crude oil and interference targets, leading to reliable data collection. The experiments simulated complex seafloor environments by introducing various interference targets, such as grass, steel pipes, shells, stones, gravel, crabs, and silt. This realistic simulation enhanced the validity of the results, demonstrating the method’s applicability in real-world scenarios.
However, the scale of the experiments was limited to a pool with specific dimensions (2 m deep, 5 m long, and 1.8 m wide). This limited scale may not capture the full range of challenges encountered in larger and more diverse marine environments, potentially impacting the generalizability of the results. It is important to note that our experiments were conducted in a 2 m deep pool. In real-world scenarios, sunken oil may be found at greater depths. Increased depth could potentially affect sonar image quality due to factors such as signal attenuation, acoustic scattering, and pressure effects. These factors might influence the performance of our identification method. Therefore, future research is crucial to investigate the impact of water depth on identification accuracy, using simulations and, where possible, field experiments at various depths. This will help to establish the robustness and limitations of our method across a range of real-world conditions and pave the way for further advancements in the field.

5. Conclusions

This article proposes a systematic framework based on the CNN algorithm to automatically identify sunken oil targets using sonar detection images. The developed framework consists of two phases. In Stage I, sonar detection images are preprocessed to extract suspected sunken oil targets. In Stage II, the CNN architecture is developed to identify all suspected sunken oil targets. A sunken oil simulation experiment is conducted in an outdoor swimming pool to collect sonar detection data. Case studies on the experimental data demonstrate the effectiveness of the presented framework. The results show the following:
(1) The proposed CNN method can effectively classify the sunken oil target, background, and interference target, which further ensures the accuracy and reliability of the proposed identification framework.
(2) Beyond the strong performance of the CNN model, the sunken oil identification framework successfully identified 83.33% of the sunken oil targets in given test images. The framework outperformed commonly used oil detection scheme SVMs, which further proves its reliability for sunken oil identification.
(3) The proposed method can provide key information such as the location of the sunken oil, which is useful for setting a basis for emergency decision-making and marine oil spill accident disposal. This paper’s work improves the effectiveness and efficiency of the existing way to detect and remove sunken oil, which is essential for marine environmental safety and the provision of marine pollution.

Author Contributions

Conceptualization, J.C. and M.G.; Data Curation, Y.Z.; Formal Analysis, N.L.; Funding Acquisition, M.G. and H.H.; Investigation, P.L.; Methodology, J.G.; Resources, J.G.; Software, N.L.; Supervision, M.G.; Validation, H.H. and Y.Z.; Writing—Original Draft, J.C. and J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by (i) the National Science Fund for Distinguished Young Scholars: Theory and Application of Deep In situ Rock Mechanics (52225403); (ii) the National Key Research and Development Program of China: Research and Application of Key Technologies for High-Precision Measurement of Reservoir Geological Parameters (2023YFF0615400); and (iii) the Shandong Provincial Natural Science Foundation (ZR2021QE059).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding authors.

Acknowledgments

The authors would like to thank Jianwei Li and CNOOC Energy Development Co., Ltd. for providing Figure 7 and Figure 8 for this study.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Zhang, X.; Mei, Z. Evaluation of Emergency Response Drills for Ship Oil Spill Based on Bayesian Network. Ship Ocean. Eng. 2018, 47, 104–108, (In Chinese with English abstract). [Google Scholar]
  2. Bao, M.; Pi, Y.; Sun, P.; Li, Y. Research Progress on “Deepwater Horizon” oil Spill Gulf of Mexico. Period. Ocean. Univ. China 2015, 45, 55–62, (In Chinese with English abstract). [Google Scholar] [CrossRef]
  3. Yuzui, T.; Kaneko, F. Frequency versus Tons of Oil Spilt Curve of Oil Tankers Using an Enhanced Power-Law Distribution Function. J. Mar. Sci. Technol. 2020, 25, 563–572. [Google Scholar] [CrossRef]
  4. Liu, P.; Liu, Y.; Cai, B.; Wu, X.; Wang, K.; Wei, X.; Xin, C. A Dynamic Bayesian Network Based Methodology for Fault Diagnosis of Subsea Christmas Tree. Appl. Ocean. Res. 2020, 94, 101990. [Google Scholar] [CrossRef]
  5. Michel, J. Spills of Nonfloating Oil: Evaluation of Response Technologies. In Proceedings of the International Oil Spill Conference, Savannah, GA, USA, 4–8 May 2008; American Petroleum Institute: Washington, DC, USA, 2008; Volume 2008, pp. 261–267. [Google Scholar]
  6. Hansen, K.A. Research Efforts for Detection and Recovery of Submerged Oil. In Proceedings of the 33rd Arctic and Marine Oil Spill Program, Halifax, NS, Canada, 7–9 June 2010; Technical Seminar; Environment Canada: Ottawa, ON, Canada, 2010; pp. 1055–1069. [Google Scholar]
  7. Hansen, K.A. Submerged Oil Response. Coast Guard. J. Saf. Secur. Sea 2013, 70, 21–23. [Google Scholar]
  8. Szafrańska, M.; Gil, M.; Nowak, J. Toward Monitoring and Estimating the Size of the HFO-Contaminated Seabed around a Shipwreck Using MBES Backscatter Data. Mar. Pollut. Bull. 2021, 171, 112747. [Google Scholar] [CrossRef] [PubMed]
  9. Lu, L.; Goerlandt, F.; Tabri, K.; Höglund, A.; Banda, O.A.V.; Kujala, P. Critical Aspects for Collision Induced Oil Spill Response and Recovery System in Ice Conditions: A Model-Based Analysis. J. Loss Prev. Process Ind. 2020, 66, 104198. [Google Scholar] [CrossRef]
  10. Sawyer, M.; Schweitzer, G.; Davis, A.; Elliott, J.; Mauseth, G.; Scott, T.M. T/B APEX 3508: Best Practices for Detection and Recovery of Sunken Oil. Int. Oil Spill Conf. Proc. 2017, 2017, 134–155. [Google Scholar] [CrossRef]
  11. Snellen, M.; Gaida, T.C.; Koop, L.; Alevizos, E.; Simons, D.G. Performance of Multibeam Echosounder Backscatter-Based Classification for Monitoring Sediment Distributions Using Multitemporal Large-Scale Ocean Data Sets. IEEE J. Ocean. Eng. 2018, 44, 142–155. [Google Scholar] [CrossRef]
  12. Tang, Q.; Liu, X.; Ji, X.; Li, J.; Chen, Y.; Lu, B. Using Seabed Acoustic Imagery to Characterize and Classify Seabed Sediment Types in the Pockmark Area of the North Yellow Sea, China. Appl. Acoust. 2021, 174, 107748. [Google Scholar] [CrossRef]
  13. Kawai, T.; Kawamura, Y.; Okada, T.; Mitsuyuki, T.; Chen, X. Sea State Estimation Using Monitoring Data by Convolutional Neural Network (CNN). J. Mar. Sci. Technol. 2021, 26, 947–962. [Google Scholar] [CrossRef]
  14. Wang, X.; Jiao, J.; Yin, J.; Zhao, W.; Han, X.; Sun, B. Underwater Sonar Image Classification Using Adaptive Weights Convolutional Neural Network. Appl. Acoust. 2019, 146, 145–154. [Google Scholar] [CrossRef]
  15. Williams, D.P. Underwater Target Classification in Synthetic Aperture Sonar Imagery Using Deep Convolutional Neural Networks. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 2497–2502. [Google Scholar]
  16. Bondarenko, O.; Kitagawa, Y. Application of the Kalman Filtering Technique for Nonlinear State Estimation in Propulsion System. J. Mar. Sci. Technol. 2021, 26, 618–631. [Google Scholar] [CrossRef]
  17. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive Histogram Equalization and Its Variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  18. Gasull, A.; Fàbregas, X.; Jiménez, J.; Marqués, F.; Moreno, V.; Herrero, M.A. Oil Spills Detection in SAR Images Using Mathematical Morphology. In Proceedings of the 2002 11th European Signal Processing Conference, Toulouse, France, 3–6 September 2002; IEEE: Piscataway, NJ, USA, 2002; pp. 1–4. [Google Scholar]
  19. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  20. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  21. Zhao, Y.; Zhang, B.; Gao, D. Construction of Petrochemical Knowledge Graph Based on Deep Learning. J. Loss Prev. Process Ind. 2022, 76, 104736. [Google Scholar] [CrossRef]
  22. Han, T.; Liu, C.; Yang, W.; Jiang, D. A Novel Adversarial Learning Framework in Deep Convolutional Neural Network for Intelligent Diagnosis of Mechanical Faults. Knowl. Based Syst. 2019, 165, 474–487. [Google Scholar] [CrossRef]
  23. Dao, P.B.; Staszewski, W.J.; Barszcz, T.; Uhl, T. Condition Monitoring and Fault Detection in Wind Turbines Based on Cointegration Analysis of SCADA Data. Renew. Energy 2018, 116, 107–122. [Google Scholar] [CrossRef]
  24. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.4842. [Google Scholar]
  25. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  26. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional Neural Networks: An Overview and Application in Radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed]
  27. Bridle, J.S. Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition. In Neurocomputing: Algorithms, Architectures and Applications; Soulié, F.F., Hérault, J., Eds.; Springer: Berlin/Heidelberg, Germany, 1990; pp. 227–236. [Google Scholar]
  28. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  29. Matkan, A.; Hajeb, M.; Azarakhsh, Z. Oil Spill Detection from SAR Image Using SVM Based Classification. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, 40, 55–60. [Google Scholar] [CrossRef]
  30. Fan, J.C.; Zhang, F.S.; Zhao, D.Z.; Wang, J. Oil Spill Monitoring Based on SAR Remote Sensing Imagery. Aquat. Procedia 2015, 3, 112–118. [Google Scholar] [CrossRef]
  31. Vapnik, V.N. An Overview of Statistical Learning Theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [PubMed]
  32. Everingham, M.; Eslami, S.M.A.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes Challenge: A Retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  33. Flach, P.; Kull, M. Precision-Recall-Gain Curves: PR Analysis Done Right. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Glasgow, UK, 2015; Volume 28. [Google Scholar]
  34. Khaldi, B.; Aiadi, O.; Kherfi, M.L. Combining Colour and Grey-Level Co-Occurrence Matrix Features: A Comparative Study. IET Image Process. 2019, 13, 1401–1410. [Google Scholar] [CrossRef]
Figure 1. Fusion image enhancement using the CNN method.
Figure 1. Fusion image enhancement using the CNN method.
Sustainability 16 06665 g001
Figure 2. Gaussian filtering. I–IV are identified targets from Gaussian filtering results: (a) original image; (b) filtered image.
Figure 2. Gaussian filtering. I–IV are identified targets from Gaussian filtering results: (a) original image; (b) filtered image.
Sustainability 16 06665 g002
Figure 3. Histogram equalization.
Figure 3. Histogram equalization.
Sustainability 16 06665 g003
Figure 4. Binary image result after applying threshold segmentation.
Figure 4. Binary image result after applying threshold segmentation.
Sustainability 16 06665 g004
Figure 5. Extraction of the edge of the target area.
Figure 5. Extraction of the edge of the target area.
Sustainability 16 06665 g005
Figure 6. Extraction of the target area.
Figure 6. Extraction of the target area.
Sustainability 16 06665 g006
Figure 7. Schematic diagram of experimental setup.
Figure 7. Schematic diagram of experimental setup.
Sustainability 16 06665 g007
Figure 8. Sonar detection image.
Figure 8. Sonar detection image.
Sustainability 16 06665 g008
Figure 9. Illustrative examples of cropped image samples. (a) Submerged oil; (b) interference targets; (c) background.
Figure 9. Illustrative examples of cropped image samples. (a) Submerged oil; (b) interference targets; (c) background.
Sustainability 16 06665 g009
Figure 10. Loss curves of training and validation sets during training.
Figure 10. Loss curves of training and validation sets during training.
Sustainability 16 06665 g010
Figure 11. Comparison of filters (a) before training; (b) after training.
Figure 11. Comparison of filters (a) before training; (b) after training.
Sustainability 16 06665 g011
Figure 12. Precision and recall rates for each category.
Figure 12. Precision and recall rates for each category.
Sustainability 16 06665 g012
Figure 13. Examples of identification results for test images. Targets marked in red boxes are identified sunken oil. Targets marked in yellow boxes are false alarms. Digits 1-4 are target numbers. (a) Original images; (b) identification results.
Figure 13. Examples of identification results for test images. Targets marked in red boxes are identified sunken oil. Targets marked in yellow boxes are false alarms. Digits 1-4 are target numbers. (a) Original images; (b) identification results.
Sustainability 16 06665 g013
Table 1. Structure of the CNN model.
Table 1. Structure of the CNN model.
Layer NameFilter NumberFilter SizeActivation
Convolutional layer55 × 5ReLU
Pooling layer-2 × 2-
Fully connected layer30-ReLU
Output3-Softmax
Table 2. Technical details of BlueView M900-2250 imaging sonar.
Table 2. Technical details of BlueView M900-2250 imaging sonar.
DeviceBlueView M900-2250
Number of beams256
Beam width1° × 20°
Pulse width-
Frequency900 kHz/2.25 MHz
Distance resolution2.54 cm
Table 3. Physical and chemical properties of the crude oil.
Table 3. Physical and chemical properties of the crude oil.
PropertiesValue
Oil density, g/m3 0.96
Oil viscosity, mm2/s 4531
Interfacial tension of the oil–seawater, mN/m25.9
Interfacial tension of the oil–freshwater, mN/m17.5
Table 4. Settings for interference targets arrangement.
Table 4. Settings for interference targets arrangement.
TypeParameter
Matrix soilSand, silty soil, sandy clay
Oil mixing rate60%, 70%, 80%
Sunken oil target shapesSquare, round, polygon, random shape
Side length or diameter5 cm, 10 cm, 20 cm, 30 cm
InterferencesGrass, steel pipes, shells, stones, gravel, crabs, and silt
Table 5. Confusion matrix of CNN-predicted results.
Table 5. Confusion matrix of CNN-predicted results.
Prediction ResultsBackgroundInterference TargetSunken Oil Target
Original Samples
Background14741
Interference target21481
Sunken oil target04197
Table 6. Precision and recall rates for each category.
Table 6. Precision and recall rates for each category.
Evaluation IndexBackgroundInterference
Target
Sunken Oil
Target
Average
P (%)98.6694.8798.9997.51
R (%)96.7198.0198.0197.58
F 1   (%)97.6896.4198.5097.53
Table 7. Sunken oil identification performance.
Table 7. Sunken oil identification performance.
ItemCNNGray-SVMGLCM-SVM
Test images222222
Actual sunken oil targets
(ground truth)
727272
Sunken oil targets identified
(true positives)
605854
Sunken oil targets unidentified
(false negatives)
121418
False alarms161813
Precision (%)78.9576.3280.60
Recall (%)83.3380.5675.00
F1 (%)81.0878.3877.70
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, J.; Gao, M.; Guo, J.; Hao, H.; Zhang, Y.; Liu, P.; Li, N. Automatic Identification of Sunken Oil in Homogeneous Information Perturbed Environment through Fusion Image Enhancement with Convolutional Neural Network. Sustainability 2024, 16, 6665. https://doi.org/10.3390/su16156665

AMA Style

Cao J, Gao M, Guo J, Hao H, Zhang Y, Liu P, Li N. Automatic Identification of Sunken Oil in Homogeneous Information Perturbed Environment through Fusion Image Enhancement with Convolutional Neural Network. Sustainability. 2024; 16(15):6665. https://doi.org/10.3390/su16156665

Chicago/Turabian Style

Cao, Jinfeng, Mingzhong Gao, Jihong Guo, Haichun Hao, Yongjun Zhang, Peng Liu, and Nan Li. 2024. "Automatic Identification of Sunken Oil in Homogeneous Information Perturbed Environment through Fusion Image Enhancement with Convolutional Neural Network" Sustainability 16, no. 15: 6665. https://doi.org/10.3390/su16156665

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop