Next Article in Journal
Physics-Based Simulation of Soft-Body Deformation Using RGB-D Data
Previous Article in Journal
Image Segmentation and Quantification of Droplet dPCR Based on Thermal Bubble Printing Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MulTNet: A Multi-Scale Transformer Network for Marine Image Segmentation toward Fishing

State Key Laboratory of Mechanical Transmission, Chongqing University, Chongqing 400044, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7224; https://doi.org/10.3390/s22197224
Submission received: 16 August 2022 / Revised: 18 September 2022 / Accepted: 19 September 2022 / Published: 23 September 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Image segmentation plays an important role in the sensing systems of autonomous underwater vehicles for fishing. Via accurately perceiving the marine organisms and surrounding environment, the automatic catch of marine products can be implemented. However, existing segmentation methods cannot precisely segment marine animals due to the low quality and complex shapes of collected marine images in the underwater situation. A novel multi-scale transformer network (MulTNet) is proposed for improving the segmentation accuracy of marine animals, and it simultaneously possesses the merits of a convolutional neural network (CNN) and a transformer. To alleviate the computational burden of the proposed network, a dimensionality reduction CNN module (DRCM) based on progressive downsampling is first designed to fully extract the low-level features, and then they are fed into a proposed multi-scale transformer module (MTM). For capturing the rich contextural information from different subregions and scales, four parallel small-scale encoder layers with different heads are constructed, and then they are combined with a large-scale transformer layer to form a multi-scale transformer module. The comparative results demonstrate MulTNet outperforms the existing advanced image segmentation networks, with MIOU improvements of 0.76% in the marine animal dataset and 0.29% in the ISIC 2018 dataset. Consequently, the proposed method has important application value for segmenting underwater images.

1. Introduction

With the development of marine ranching, there are serious drawbacks to traditional manual fishing, including the high risk of fishing, danger of fishermen, low efficiency, and small working area. Therefore, it is worth researching the autonomous underwater vehicle (AUV). The object segmentation plays an important role in the AUV. In recent years, machine vision technology has been widely used for target detection and segmentation owing to the advantages of low cost and high precision. Some methods based on deep learning [1,2,3] have been proposed for the detection and recognition of marine organisms. However, these methods based on convolutional neural networks (CNNs) mainly focus on image classification tasks, and they cannot predict the specific shapes and sizes of marine organisms, which affects the accurate grab of marine animals by the robotic gripper. Consequently, it is extremely significant for implementing accurate marine image segmentation.
As an important branch of computer vision, image semantic segmentation can perform pixel-level classification by labelling each pixel of an image into a corresponding class, which is also known as dense prediction. Semantic segmentation is suitable for a variety of tasks, including autonomous vehicles and medical image diagnostics. However, there are few studies on marine animal segmentation. In addition, due to the noise caused by flowing particles, light scattering, and a complex underwater background, the acquired marine images are generally vague or of poor quality, which seriously affects the accuracy of marine image segmentation. To overcome these issues mentioned above, this work aims to explore a new transformer-based image segmentation method for improving the precision of marine animal segmentation and meeting the requirements of automatic fishing.
In the past decade, deep learning has become a hot topic in image segmentation. In particular, various CNNs have been widely applied to semantic segmentation. Although the CNN network does well in extracting the low-level features, it is difficult to capture the global semantic information owing to the limitation of the receptive fields of convolution kernels. On the other hand, as a new technique of deep learning, the transformer has excellent performance in dealing with high-level semantic information, which can effectively capture the long-distance dependence. However, since the patch size is fixed, it is hard for the transformer to acquire the low-resolution and multi-scale feature maps, which brings great difficulties to the segmentation task. Whereupon we developed a new multi-scale transformer network for semantic segmentation that possesses the advantages of both CNN in extracting the low-level features and transformer in processing the relationship between two visual elements, and it is named MulTNet. The MulTNet mainly consists of a dimensionality reduction CNN module, a multi-scale transformer module, and a decoder. With the principle of progressive downsampling, the dimensionality reduction module is designed. It first uses two convolution layers to gradually decrease the resolution of feature maps while flexibly increasing the number of channels, and the feature maps with sufficient low-level semantic information are obtained, then they are flattened into a low dimension matrix as the input of the multi-scale transformer module. The multi-scale transformer module is constructed through four parallel small-scale encoder layers with different heads and a conventional large-scale encoder layer. In this module, the four small-scale encoder layers can help to extract the multi-scale contextural features and improve the computational efficiency through parallel computing, while the large-scale transformer encoder layer can further extract the high-level features on the basis of multi-scale features. Finally, several CNN-based decoder layers are used to recover the original image resolution for better segmentation results. Compared with other existing typical segmentation methods, MulTNet demonstrates higher segmentation performance in the marine animal dataset and the public ISIC 2018 dataset. Therefore, it has great application potential in marine fishing. The contributions of this paper are summarized as:
(1) A new semantic segmentation network named MulTNet is proposed, and it can extract the low-level features and the multi-scale high-level semantic features, thereby improving the segmentation accuracy. The proposed network is suitable for various object segmentation tasks, especially for marine animal segmentation;
(2) A new multi-scale transformer module has been developed. The four transformer layers with different heads can parallel extract the contextural characteristics in different subspaces and scales, which enhances the feature expression ability of the proposed network and increases the computation speed;
(3) A dimensionality reduction CNN module based on progressive downsampling is designed for extracting the low-level features. Via the flatten operation, the obtained low-dimension matrix can alleviate the computational burden of the subsequent transformer. Additionally, the ablation experiment verifies the effectiveness of this module.
The remainder of this paper is organized as follows. A summary of the related work is reviewed in Section 2 and the overall architecture of our proposed method is described in detail in Section 3. Section 4 presents the training details and discusses the results with comparisons to state-of-the-art methods. Section 5 focuses on conclusions and the prospect of future work.

2. Related Work

Semantic segmentation: As one of the earliest segmentation networks based on the encoder-decoder structure, FCN [4] designed a skip architecture to combine the deep and coarse information with the shallow and fine information to improve the segmentation accuracy. However, FCN does not consider global context information, and its predictions are relatively rough [5]. As one of the most popular methods, U-Net [6] can utilize the global location and context simultaneously, which has good performance in image segmentation [7]. As an extension to U-Net, Attention U-Net [8] showed great improvements in the accuracy and sensitivity of the foreground pixels by integrating a self-attention gating module into the conventional CNN-based segmentation network. Lin et al. [9] developed RefineNet to learn the high-level semantic features by adding residual connections. The authors of [10] used ResUNet take advantage of both the U-Net network and deep residual learning and promoted information propagation by concatenating the low-level details and the high-level semantic features.
Furthermore, research by [11,12,13] addressed expanding receptive fields and improving multi-scale information for semantic segmentation. Yu et al. [14] and Chen et al. [15] put forward dilated or atrous convolutions. The authors of [16] made use of Deeplabv3+ and used atrous separable convolution in both the decoder module and the ASPP model to probe different features from different scales and spaces by applying depth-dependent separable convolution to ASPP. On the basis of dilated residual networks, the authors of [17] utilized PSPNet and designed the PPM module to enhance the extraction ability of context information from different subspaces. As a cascading knowledge diffusion network, the authors of [18] utilized CKDNet and boosted segmentation performance by fusing information learnt from different tasks. Although the above approaches have great success in enriching the spatial context, it is still hard to capture the long-range dependencies and global information.
Underwater segmentation: Although the current segmentation methods show excellent performance on various tasks, such as medical image segmentation and autonomous vehicle image segmentation, they face a challenge for the accurate segmentation of underwater images. Hence, some specific methods were proposed to segment the marine animal images. Wei et al. [19] designed the Unet-FS network to segment different types of free surfaces. Yao et al. [20] proposed a new segmentation method for fish images, and it used the mathematical morphology and K-means clustering segmentation algorithm to separate the fish from the background. The authors of [21,22,23] explored several non-End2End segmentation architectures to obtain high accuracy on the seagrass images. Furthermore, a number of underwater image enhancement algorithms [24,25,26,27,28] have been proposed, which can be beneficial to the subsequent segmentation. If orthogonal polynomials are used [29,30], the accuracy of segmentation can be further improved. To improve the quality of images, some advanced underwater image sensors have been proposed [31,32,33]. Several publicly available underwater datasets [34,35,36] have been published to test the performance of various image processing approaches. In this paper, we focus on researching a new marine animal segmentation method to enhance the ability of automatic fishing.
Transformer: With the self-attention mechanism and Seq2Seq structure, the transformer was originally designed for processing natural language. Vaswani et al. [37] first proposed a transformer for machine translation and achieved good results. Devlin et al. [38] proposed a new language representation model called BERT to pretrain a transformer based on the unlabeled text considering the content of each word. BERT demonstrated excellent performance on eleven NLP tasks. Other transformer variants have also been proposed for solving NLP tasks, such as the famous BART [39], GPT-3 [40], and T5 [41]. Recently, due to the strong representation capabilities of transformers, they have been widely used for the various vision tasks, including semantic segmentation [42,43], object detection [44,45], and image recognition [46]. Dosovitskiy et al. [47] proposed a vision transformer (ViT), and it possessed the satisfactory performance of image classification owing to the large-scale pretraining of a convolution-free transformer. As ViT requires a large amount of pretraining data, Touvron et al. [48] developed a data-efficient image transformer (DeiT), and it utilized a token-based distillation strategy and teacher model to guide the better learning of DeiT. A swin transformer [49] was built as a vision backbone using hierarchical architecture and shifted window operation.
Although these transformer-based methods have achieved great success in image classification tasks, they rarely obtain the feature maps with different scales, which affects their segmentation capacities. Thereupon, some hybrid models combining CNNs, and transformers were proposed, especially for small-scale datasets. Chen et al. put forward TransUnet for medical image segmentation via combining transformer with U-Net. Xie et al. [50] proposed a hybrid network (CoTr) that effectively bridged CNNs and transformers, achieving a significant performance improvement for 3D medical image segmentation. Incorporating with a transformer, TrSeg [51] adaptively produced multi-scale contextual information with dependencies on original contextual information. As a task-oriented model based on skin lesion segmentation, FAT-Net [52] increased the receptive field for extracting more global context information.
Unfortunately, the above hybrid networks just use the transformers but do not improve the feature extraction ability of the transformer. Meanwhile, the transformer usually has a large computational burden. To address the above issues, a multi-scale transformer is explored in this work, and it can not only effectively extract the multi-scale features but also increase the computing speed. MulTNet, which integrates CNN and the multi-scale transformer, is also proposed to combine the ability of CNN in capturing local visual perception with the capacity of the transformer in building long-range dependencies. The low-level features and the multi-scale high-level semantic features can finely represent the characteristic information from various objects and noise. The proposed network has powerful segmentation performance, and it is superior to the current classical segmentation methods, which is validated by two datasets. All the segmentation method mentioned in the literature are listed in Table 1.

3. Proposed Network

The proposed MulTNet has a network architecture of encoder-decoder. To reduce the computational burden, a dimensionality reduction module based on CNN is designed, which consists of two convolution layers and a flattening layer. Two convolutional layers can effectively reduce the dimensions of feature maps, which is helpful for simplifying the tensor operation and reducing the memory requirement. Then, we propose a multi-scale transformer module as the encoder of MulTNet. The encoder consists of four parallel small-scale transformer encoder layers and one conventional large-scale transformer encoder layer. Four small-scale transformer encoder layers with different heads can extract useful information from different scales, and their outputs are combined into the input of the large-scale conventional transformer layer for further feature extraction. Via combining both transformer and CNN branches in our network, not only can the abundant local features be extracted, but also the important global context information can be fully captured for marine animal segmentation. The decoder uses step-by-step convolutions to implement the restoration of the segmented image. The detailed structure diagram of MulTNet is shown in Figure 1.

3.1. Dimensionality Reduction CNN Module

In most current vision-transformer networks, the convolution operation is adopted for downsampling, that is, the front-end high-dimensional image is transformed into a low-dimensional tensor. As previously mentioned, a hybrid model combining transformer with CNN always shows better performance than a single transformer. Furthermore, convolutional layers generate the feature maps pixel by pixel [53], which contributes to an excellent performance in capturing local features. It then follows that step-by-step downsampling is employed to reduce the resolution of input images. Simultaneously, the channels in each CNN layer are gradually expanded. In this way, the local information of the input image can be extracted as much as possible. Specifically, we first reduce the resolution and expand the channels of the original image x R H × W × 3 is converted into X R H / 16 × W / 16 × C by adopting two convolution layers, where H, W, and C denote the height, width, and channels of the feature maps respectively. One layer has 192 convolution kernels with a size of ( 8 × 8 ) and a stride of 8, while the other layer has 768 convolution kernels with a size of ( 2 × 2 ) and a stride of 2. The two convolution operations can be expressed as:
f m 1 = C o n v 8 × 8 × 192 ( R E L U ( B a t c h N o r m ( X i n p u t ) ) )
f m 2 = C o n v 2 × 2 × 768 ( R E L U ( B a t c h N o r m ( f m 1 ) ) )
where fm1 and fm2, respectively, denote the output feature maps of the first layer and the second layer.
Considering that the transformer just accepts the 1D sequence of patch embeddings as the input, every single feature map is flattened into a 1D sequence. In order to handle different patch sizes, the linear projection is used to map the patch embeddings into the vectors with a fixed length, and they form a C-dimensional embedding space E R L × C , where L is the length of a single sequence, which is set to HW/256, and C is the number of channels. To keep the spatial position information of patches, the single position embedding is added with the patch embedding to obtain the final input of the multi-scale transformer module. This computational procedure is formulated as:
E t = p 1 + e 1 , p 2 + e 2 , , p C + e C
where pi and ei, respectively, denote the single position embedding and the 1D patch embedding in the i-th channel; E t represents the final linear flattened 2D embedding space.

3.2. Multi-Scale Transformer Module

Without the capability of building long-range relationships, CNNs cannot precisely distinguish the differences between pixels for some challenging work, including underwater image segmentation, where the problems of insufficient resolution, low contrast, and blurred marine animal boundaries are usually present. Unlike the convolutional layers, which gain global features only by stacking more layers, the transformer can effectively build the long-range relationships to capture the rich global context information. Inspired by the multi-scale strategy in deep learning [54], we employ a multi-scale transformer module to capture global features from different subregions and scales. A conventional transformer struggles to capture contextural information in different subregions and scales with a fixed number of heads and too many patches for training, and its computational efficiency is low. The 2D embedding space, E t obtained by the dimension reduction CNN module is then divided evenly into four small embedding matrices, E s R L × C / 4 which are fed into four parallel transformer encoder layers with different heads.
As the core of the transformer, multi-head attention is proposed via the stacking of multiple simple self-attention functions, which helps to capture different spatial and semantic information from different subspaces. The output of a simple self-attention module can be written as:
A T T ( Q , K , V ) = S o f t m a x ( Q K T d K ) V
where Q, K, and V represents query, key, and value matrices respectively; d K denotes the dimension of Q and K. The multi-head attention linearly projects Q, K, and V into h different subspaces via the self-attention operation, and then the outputs of multiple heads (self-attention modules) are concatenated. The calculation process of multi-head attention is formulated as:
H i = A T T ( Q W i Q , K W i K , V W i V )
A T T M u l t i H e a d ( Q , K , V ) = C o n c a t ( H 1 , , H h ) W o
where Q W i K , K W i K , and V W i V respectively represent the query, key, and value of the i-th head, and W i Q , W i K R D m o d e l × d k and W i V R D m o d e l × d v are the corresponding weight matrices, in which D m o d e l is the dimension of embedding and d v denotes the dimension of Q. W O R h d v × D m o d e l denotes the weight matrix.
In our study, four transformer encoder layers with 1, 2, 3, and 4 heads, respectively, are designed to capture more contextural information from different scales and different subregions. Through a series of tests, we find that the semantic features at different scales can be well extracted according to the designed head numbers. Each transformer layer takes a small embedding matrix ( 256 × 192 ) as the input, and the four transformer layers can be formulated as:
T s _ 1 = A T T 1 M u l t i H e a d ( Q 1 , K 1 , V 1 ) = A T T ( Q W 1 Q , K W 1 K , V W 1 V ) N 1
T s _ 2 = A T T 2 M u l t i H e a d ( Q 2 , K 2 , V 2 ) = C o n c a t ( H 1 2 , H 2 2 ) N 2
T s _ 3 = A T T 3 M u l t i H e a d ( Q 3 , K 3 , V 3 ) = C o n c a t ( H 1 3 , H 2 3 , H 3 3 ) N 3
T s _ 4 = A T T 4 M u l t i H e a d ( Q 4 , K 4 , V 4 ) = C o n c a t ( H 1 4 , H 2 4 , H 3 4 , H 4 4 ) N 4
where N 1 , N 2 , N 3 , N 4 R ( h C 4 h ) × C 4 denote the weight matrices obtained in the first, second, third, and fourth transformer layers, respectively; H j i R C 4 i denotes the j-th head of the ith transformer layer. We concatenate all these outputs into a R L × C matrix, which is written as:
M u l T = C o n c a t ( T 1 , T 2 , T 3 , T 4 )
To better establish the long-distance dependencies among those results obtained by the four small-scale transformer encoder layers, we adopt another large-scale transformer encoder layer with 5 heads to further capture the global contextural characteristics, and the computational process is formulated as:
T o u t p u t = A T T M u l t i H e a d ( Q o , K o , V o ) = C o n c a t ( H 1 , H 2 , H 3 , H 4 , H 5 ) N o
where N O R C × C ; H i R C 5 denotes the i-th head in the large-scale transformer obtained by the even division of MulT.

3.3. Loss of MulTNet

The categorical cross-entropy loss (CE loss) function is expressed as the difference between the true probability distribution and the predicted probability distribution. The smaller the value of cross entropy, the better the prediction effect of the model. The formula can be expressed as:
E = i = 1 N y i l o g ( S ( f ( x i ) ) )
where N denotes the number of categories; y i denotes the true target tensor for the i-th category of samples, while f ( x i ) denotes the predicted target tensor, that is, the output of the decoder, as shown in Figure 1.; S represents the softmax function.

4. Experiments and Discussion

4.1. Dataset Description

Marine animal dataset: The underwater images come from a Real-world Underwater Image Enhancement (RUIE) dataset [55], which can be downloaded from https://github.com/dlut-dimt/Realworld-Underwater-Image-Enhancement-RUIE-Benchmark (accessed on 7 August 2021). RUIE is the first large underwater image dataset for validating various object detection algorithms. In this work, 1396 underwater images are selected for our segmentation task, which have four categories, including background, sea cucumber, sea urchin, and starfish. LabelMe that can label the images is utilized to obtain the ground truth of the dataset. Then the dataset is divided into the training set, validation set, and test set, which respectively have 1116, 140, and 140 samples according to the ratio of 8:1:1. These images have various light scattering effects for verifying the proposed segmentation method.
ISIC dataset: This dataset was published by the International Skin Imaging Collaboration (ISIC) in 2018. As a large-scale dataset of dermoscopic images, it contains 2594 images with their corresponding ground truth annotations, which are available at https://challenge2018.isic-archive.com/ (accessed on 19 October 2021). For the ISIC dataset, we used 2074 images as the training set, 260 images as the test set, and 260 images as the validation set.

4.2. Model Trainings

We train MulTNet using mini-batch stochastic gradient descent (SGD) with an initial learning rate of 0.03, batch size of 6, and momentum of 0.9. To reduce the training time, the multi-scale transformer module adopts parallel training. For the marine animal dataset, 400 epochs are implemented, while 200 epochs are implemented for the ISIC dataset. Other hyper-parameters for both datasets are listed in Table 2. Furthermore, the following software and hardware configurations were used in the experiments: TensorFlow2.4 + Windows10 + CUDA 10.2 + CUDNN7.6.5 + NVIDIA GeForce RTX 3080.

4.3. Evaluation Metrics

In order to quantitatively evaluate the segmentation capability of the proposed MulTNet, we use a variety of evaluation metrics, including MIOU, IOU, Acc, and MPA [56]. The confusion matrix between prediction and ground truth is calculated via the number of positive pixels for the true identification (TP), the number of negative pixels for the true identification (TN), the number of negative pixels for the false identification (FP), and the number of positive pixels for the false identification (FN). Assume that i represents the true value, j represents the predicted value, p i j denotes the number of pixels originally belonging to class i but predicted to be class j, and k + 1 denotes the number of categories (including background).
IOU denotes the overlap rate of the predicted images and ground truths, which is the ratio between their intersection and union. The ideal situation is a complete overlap, namely, the ratio is equal to 1. IOU is defined as
I O U = T P T P + F N + F P
MIOU is obtained by calculating the average of IOUs for all categories, and the computational formula is written as:
M I O U = 1 k + 1 i = 0 k p i i j = 0 k p i j + j = 0 k ( p i j p i i )
Acc represents the pixel accuracy rate, and it is calculated as follows:
P A = i = 0 k p i i i = 0 k j = 0 k p i j
MPA stands for mean pixel accuracy, which is defined as follows:
M P A = 1 k + 1 i = 0 k p i i j = 0 k p i j

4.4. Comparison with State-Of-The-Art Methods

Using the marine animal dataset and the ISIC dataset, the proposed MulTNet is compared with the state-of-the-art methods including FCN, U-Net, Attention U-Net, RefineNet, ResUnet, DeeplabV3+, CKDNet, PSPNet, TrSeg, SegFormer [57], and FAT-Net. The evaluation metrics defined in Section 4.3 are used for quantitatively analyzing the segmentation abilities of various networks. Meanwhile, for a fair comparison, the hyper-parameters of the comparative methods are the same as those of MulTNet, which are listed in Table 2.
Firstly, all the used segmentation networks are applied to the marine animal dataset, which has great difficulty in segmentation owing to the blurry images and irregular shapes of marine organisms in the complex marine environment. The evaluation metrics obtained by MulTNet and the comparative networks are listed in Table 3. It is obvious that the MPA, MIOU, and Acc of MulTNet are up to 47.69%, 45.63%, and 97.27%, respectively, which are 0.52%, 0.76%, and 0.11% higher than those of the second-ranked FAT-Net. In the meantime, they are 10.82%, 10.15%, and 0.22% higher than those of the lowest-ranked FCN. In addition, the proposed network has the largest IOUs for the objects of sea urchin and starfish, while the IOU of MulTNet is only smaller than that of PSPNet for the object of sea cucumber. Consequently, our proposed network has better segmentation performance for the marine animal dataset than other methods.
FCN, U-Net, and Attention U-Net cannot make accurate predictions of marine animals when the underwater images are mostly blurry and have low contrast because they lack strong extraction ability for global contextual information. By taking advantage of both the U-Net framework and residual units, ResUnet obtains better accuracy performance than RefineNet. Relying on atrous convolution to obtain a large receptive field and the ASPP module to encode multi-scale contextual information, Deeplabv3+ also has a higher performance than ResUNet. CKDNet exploits two novel features entangled to boost the performance of classification and fine-level segmentation, so it outperforms the Deeplabv3+ method. Compared to CKDNet, PSPNet utilizes the pyramid pooling module to extract global context information from different regions, which shows the improved segmentation performance. PSPNet generates multi-scale contextual information without any prior information, while TrSeg can better generate it by adapting a transformer architecture to build dependencies on original contextual information. Without the interpolation of positional codes leading to decreased performance, SegFormer employs a MLP decoder to aggregate features from different layers, hence it is superior to TrSeg. As a hybrid encoder-decoder network consisting of CNN and transformer encoder branches, Fat-Net uses a FAM module and fuses it into each skip connection for enhancing the feature fusion between two feature maps, which can enhance the segmentation performance. However, these methods cannot establish long-term relationships or extract global context information, preventing segmentation performance from being improved further. Our proposed method, which is based on DRCM and MTM, can fully capture both local features and establish long-range dependencies to improve segmentation performance.
To visualize the segmentation results obtained by various networks, several typical examples of segmentation results are shown in Figure 2. The first and second columns are the original images and ground truths, respectively, and the next columns are, respectively, the segmentation results of FCN, U-Net, Attention U-Net, RefineNet, ResUnet, DeeplabV3+, PSPNet, TrSeg, CKDNet, SegFormer, FAT-Net, and MulTNet. Figure 2 shows that the predicted shapes of sea urchins and starfish obtained by MulTNet are more similar to the ground truths, particularly for some irregular or inconspicuous contours. The intuitive comparison further demonstrates that MulTNet has a higher accuracy of segmentation.
In addition, several examples from different levels of blur are selected to demonstrate the segmentation ability of our proposed method. As shown in Figure 3, our proposed method can precisely segment the marine animal nearly without any visible defects in the blurred underwater situation, primarily because DRCM with a large number of channels can precisely capture abundant low-level information. Meanwhile, the MTM can establish the long-distance dependence between blurred pixels and fully extract the high-level contextural information from different subregions and scales.
Next, to further verify the superiority of MulTNet, all methods are applied to segment the public ISIC 2018 dataset, and Table 4 shows their evaluation results. Evidently, four evaluation metrics of the proposed MulTNet outperform those of the existing classical networks, and the MPA, IOU, MIOU, and Acc of MulTNet are up to 94.04%, 81.48%, 88.09%, and 95.71%, respectively. FAT-Net and CKDNet are specially designed for skin lesion segmentation. It then follows that MulTNet has stronger segmentation ability and higher pixel-level segmentation performance than other typical networks. Similarly, several typical segmentation examples obtained by all the methods are illustrated in Figure 4 for a visualized comparison. It can be noted from Figure 4. that MulTNet extracts more details related to the shapes, especially for the dermoscopic image shown in the last row. Therefore, the advantage of the proposed MulTNet is validated again.
As is well-known, the loss function is used for network training, and it can measure the prediction accuracy and convergence rate. With regard to the marine animal dataset and the ISIC 2018 dataset, the loss curves of all networks are respectively drawn in Figure 5a,b. In Figure 5a, MulTNet has a minimum converged loss of over 0.1% lower than that of second-ranked FAT-Net, which means our segmentation network can better learn the semantic features compared to other segmentation networks. Meanwhile, the convergence rate of our model is basically the same as Attention U-Net. From Figure 5b, it can be clearly observed that our proposed MulTNet has the fastest convergence rate, and its converged loss is approximately equal to that of ResUNet. Compared with other networks, our proposed method can be convergent with fewer epochs, which leads to a faster convergence speed. The above analysis shows that MulTNet has higher convergence performance than other typical segmentation networks for both datasets, so it has more powerful segmentation ability.
Finally, we compare the total parameters, training speed, and testing speed of each model, and the results are illustrated in Table 5. FCN-8s have the maximum number of parameters (134 M), which leads to the slowest training speed. However, not all of these parameters have a positive influence on the final segmentation performance. U-Net is a lightweight U-shaped network that has the minimum total parameters and the fastest training speed. Owing to the proposed dimensionality reduction CNN module and multi-scale transformer module, the training speed and testing speed of MulTNet are the second fastest, which only take 0.183 s per iteration and 0.087 s per image, respectively. Although the training speed and testing speed of our model are slightly slower than those of U-Net, the segmentation accuracy of our model is obviously higher than that of U-Net.

4.5. Ablation Experiment

To verify the effectiveness of each module of the proposed dimensionality reduction CNN module (DRCM) and multi-scale transformer module (MTM), the original transformer, DRCM-Transformer, and MulTNet are used in the ablation experiments. For the ISIC and marine animal datasets, the segmentation results obtained by three approaches are listed in Table 6. It can be seen from this table that the MIOU and MPA on the marine animal dataset obtained by the original transformer are just 23.99% and 25.00%, respectively, as it cannot capture the characteristics at different scales, which also represent the feature information from different objects and noise. Compared with the original transformer, the DRCM-Transformer extracts the low-level features, so it possesses stronger segmentation ability. By adding MTM into DRCM-Transformer, the high-level contextural characteristics in different subspaces and scales can be extracted, so MulTNet has higher MIOU and MPA than DRCM-Transformer. The ablation experiment results show that DRCM and MTM play important roles in the proposed framework, and they can enhance the segmentation performance of the original transformer.

5. Conclusions

Since the shapes of various marine animals are very different and the collected images are influenced by the complex and varying underwater environment, it is a great challenge to achieve precise marine image segmentation. To fill this gap, a new multi-scale transformer segmentation network named MulTNet is proposed. The MulTNet is composed of a dimensionality reduction CNN module, a multi-scale transformer module, and a decoder. The dimensionality reduction CNN module is designed for generating the low-level feature maps, and then they are fed into the multi-scale transformer module, which is the main contribution of this work. The proposed multi-scale transformer module utilizes four small-scale transformer layers with different heads to extract the contextual characteristics from different subregions and scales, while a large-scale transformer is designed for further extracting the high-level semantic features. In addition, the parallel computation of four small-scale layers can effectively increase the computation speed. The results of contrast experiments indicate that MulTNet has higher segmentation performance than the existing typical methods for both the marine animal dataset and the ISIC dataset. Therefore, the proposed segmentation network can be better applied to segment the marine animals. In future research, an effective image preprocessing method will be explored for enhancing the quality of the marine images, which can help improve the segmentation accuracy, and we will apply our method to the AUVs for automatic fishing.

Author Contributions

Formal analysis and writing—original draft, X.X.; methodology, conceptualization, and funding acquisition, Y.Q.; experimental and programming, D.X.; visualization, R.M.; data curation and software, J.X. All authors have read and agreed to the published version of the manuscript.

Funding

The National Natural Science Foundation of China (nos. 62033001 and 52175075), Chongqing Research Program of Basic Research, and Frontier Exploration (no. cstc2021ycjh-bgzxm0157).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

We evaluate our network on the public dataset of International Skin Imaging Collaboration (ISIC 2018). The information link is: https://challenge2018.isic-archive.com/ (accessed on 19 October 2021).

Acknowledgments

The work described in this paper was supported by the National Natural Science Foundation of China (nos. 62033001 and 52175075) and the Chongqing Research Program of Basic Research and Frontier Exploration (no. cstc2021ycjh-bgzxm0157).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, F.L.; Yao, J.; Zhu, H.; Wang, C. Marine organism detection and classification from underwater vision based on the deep CNN method. Math. Probl. Eng. 2020, 2020, 3937. [Google Scholar] [CrossRef]
  2. Zhuang, P.; Xing, L.; Liu, Y.; Guo, S.; Qiao, Y. Marine Animal Detection and Recognition with Advanced Deep Learning Models. In CLEF; Working Note; Shenzhen Institutes of Advanced Technology Chinese Academy of Sciences: Shenzhen, China, 2017. [Google Scholar]
  3. Cao, Z.; Principe, J.C.; Ouyang, B.; Dalgleish, F.; Vuorenkoski, A. Marine animal classification using combined CNN and hand-designed image features. In Proceedings of the OCEANS 2015—MTS/IEEE Washington, Washington, DC, USA, 19–22 October 2015; pp. 1–6. [Google Scholar]
  4. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  5. Xi, D.J.; Qin, Y.; Luo, J.; Pu, H.Y.; Wang, Z.W. Multipath fusion Mask R-CNN with double attention and its application into gear pitting detection. IEEE Trans. Instrum. Meas. 2021, 70, 5006011. [Google Scholar] [CrossRef]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the MICCAI 2015: Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  7. Qin, Y.; Wang, Z.; Xi, D.J. Tree CycleGAN with maximum diversity loss for image augmentation and its application into gear pitting detection. Appl. Soft Comput. 2022, 114, 108130. [Google Scholar] [CrossRef]
  8. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, preprint. arXiv:1804.03999. [Google Scholar]
  9. Lin, G.; Milan, A.; Shen, C.; Reid, I. Refinenet: Multi-path refine-ment networks for high-resolution semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1925–1934. [Google Scholar]
  10. Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef]
  11. Jin, H.; Cao, L.; Kan, X.; Sun, W.; Yao, W.; Wang, X. Coal petrography extraction approach based on multiscale mixed-attention-based residual U-net. Meas. Sci. Technol. 2022, 33, 075402. [Google Scholar] [CrossRef]
  12. Wang, Z.; Guo, J.; Huang, W.; Zhang, S. High-resolution remote sensing image semantic segmentation based on a deep feature aggregation network. Meas. Sci. Technol. 2021, 32, 095003. [Google Scholar] [CrossRef]
  13. Sang, H.W.; Zhou, Q.H.; Zhao, Y. PCANet: Pyramid convolutional attention network for semantic segmentation. Image Vis. Comput. 2020, 103, 103997. [Google Scholar] [CrossRef]
  14. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, preprint. arXiv:1511.07122. [Google Scholar]
  15. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, preprint. arXiv:1412.7062. [Google Scholar]
  16. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  17. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  18. Jin, Q.; Cui, H.; Sun, C.; Meng, Z.; Su, R. Cascade knowledge diffusion network for skin lesion diagnosis and segmentation. Appl. Soft. Comput. 2021, 99, 106881. [Google Scholar] [CrossRef]
  19. Wei, Z.; Zhai, G.; Wang, Z.; Wang, W.; Ji, S. An artificial intelligence segmentation method for recognizing the free surface in a sloshing tank. Ocean Eng. 2021, 220, 108488. [Google Scholar] [CrossRef]
  20. Yao, H.; Duan, Q.; Li, D.; Wang, J. An improved K-means clustering algorithm for fish image segmentation. Math. Comp. Modell. 2013, 58, 790–798. [Google Scholar] [CrossRef]
  21. Martin-Abadal, M.; Riutort-Ozcariz, I.; Oliver-Codina, G.; Gonzalez-Cid, Y. A deep learning solution for Posidonia oceanica seafloor habitat multiclass recognition. In Proceedings of the OCEANS 2019-Marseille, Marseille, France, 17–20 June 2019; pp. 1–7. [Google Scholar]
  22. Martin-Abadal, M.; Guerrero-Font, E.; Bonin-Font, F.; Gonzalez-Cid, Y. Deep semantic segmentation in an AUV for online posidonia oceanica meadows identification. IEEE Access 2018, 6, 60956–60967. [Google Scholar] [CrossRef]
  23. Sengupta, S.; Ersbøll, B.K.; Stockmarr, A. SeaGrassDetect: A novel method for the detection of seagrass from unlabelled underwater videos. Ecol. Inform. 2020, 57, 101083. [Google Scholar] [CrossRef]
  24. Wang, L.; Shang, F.; Kong, D. An image processing method for an explosion field fireball based on edge recursion. Meas. Sci. Technol. 2022, 33, 095021. [Google Scholar] [CrossRef]
  25. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color balance and fusion for underwater image enhancement. IEEE Trans. Image Process. 2017, 27, 379–393. [Google Scholar] [CrossRef]
  26. Iqbal, K.; Salam, R.A.; Osman, A.; Talib, A.Z. Underwater image enhancement using an integrated colour model. Int. J. Comput. Sci. 2007, 34, 239–244. [Google Scholar]
  27. Zhao, X.; Jin, T.; Qu, S. Deriving inherent optical properties from background color and underwater image enhancement. Ocean Eng. 2015, 94, 163–172. [Google Scholar] [CrossRef]
  28. Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1382–1386. [Google Scholar]
  29. Mahmmod, B.M.; Abdulhussain, S.H.; Suk, T.; Hussain, A. Fast Computation of Hahn Polynomials for High Order Moments. IEEE Access 2022, 10, 48719–48732. [Google Scholar] [CrossRef]
  30. Al-Utaibi, K.A.; Abdulhussain, S.H.; Mahmmod, B.M.; Naser, M.A.; Alsabah, M.; Sait, S.M. Reliable recurrence algorithm for high-order Krawtchouk polynomials. Entropy 2021, 23, 1162. [Google Scholar] [CrossRef] [PubMed]
  31. Skinner, K.A.; Matthew, J.-R. Underwater image dehazing with a light field camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 62–69. [Google Scholar]
  32. Bonin, F.; Burguera, A.; Oliver, G. Imaging systems for advanced underwater vehicles. J. Marit. Res. 2011, 8, 65–86. [Google Scholar]
  33. Eleftherakis, D.; Vicen-Bueno, R. Sensors to increase the security of underwater communication cables: A review of underwater monitoring sensors. Sensors 2020, 20, 737. [Google Scholar] [CrossRef] [PubMed]
  34. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE T. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed]
  35. Duarte, A.; Codevilla, F.; Gaya, J.D.O.; Botelho, S.S. A dataset to evaluate underwater image restoration methods. In Proceedings of the OCEANS 2016-Shanghai, Shanghai, China, 10–123 April 2016; pp. 1–6. [Google Scholar]
  36. Radolko, M.; Farhadifard, F.; Von Lukas, U.F. Dataset on underwater change detection. In Proceedings of the OCEANS 2016 MTS/IEEE Monterey, Monterey, CA, USA, 19–23 September 2016; pp. 1–8. [Google Scholar]
  37. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Polosukhin, I. Attention is all you need. Adv. Neural. Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  38. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, preprint. arXiv:1810.04805. [Google Scholar]
  39. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Zettlemoyer, L. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv 2019, preprint. arXiv:1910.13461. [Google Scholar]
  40. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Amodei, D. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  41. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv 2019, preprint. arXiv:1910.10683. [Google Scholar]
  42. Wang, Y.; Xu, Z.; Wang, X.; Shen, C.; Cheng, B.; Shen, H.; Xia, H. End-to-end video instance segmentation with transformers. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 8741–8750. [Google Scholar]
  43. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, preprint. arXiv:2102.04306. [Google Scholar]
  44. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  45. Beal, J.; Kim, E.; Tzeng, E.; Park, D.H.; Zhai, A.; Kislyuk, D. Toward transformer-based object detection. arXiv 2020, preprint. arXiv:2012.09958. [Google Scholar]
  46. Zhang, Q.; Yang, Y. ResT: An efficient transformer for visual recognition. arXiv 2021, preprint. arXiv:2105.13677. [Google Scholar]
  47. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Houlsby, N. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, preprint. arXiv:2010.11929. [Google Scholar]
  48. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  49. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. arXiv 2021, preprint. arXiv:2103.14030. [Google Scholar]
  50. Xie, Y.; Zhang, J.; Shen, C.; Xia, Y. CoTr: Efficiently Bridging CNN and Transformer for 3D Medical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2021; pp. 171–180. [Google Scholar]
  51. Jin, Y.; Han, D.; Ko, H. TrSeg: Transformer for semantic segmentation. Pattern Recogn. Lett. 2021, 148, 29–35. [Google Scholar] [CrossRef]
  52. Wu, H.; Chen, S.; Chen, G.; Wang, W.; Lei, B.; Wen, Z. FAT-Net: Feature adaptive transformers for automated skin lesion segmentation. Med. Image Anal. 2022, 76, 102327. [Google Scholar] [CrossRef]
  53. Qian, Q.; Qin, Y.; Wang, Y.; Liu, F. A new deep transfer learning network based on convolutional auto-encoder for mechanical fault diagnosis. Measurement 2021, 178, 109352. [Google Scholar] [CrossRef]
  54. Xiang, S.; Qin, Y.; Luo, J.; Pu, H.; Tang, B. Multicellular LSTM-based deep learning model for aero-engine remaining useful life prediction. Reliab. Eng. Syst. Saf. 2021, 216, 107927. [Google Scholar] [CrossRef]
  55. Liu, R.; Fan, S.; Zhu, M.; Hou, M.; Luo, Z. Real-world underwater enhancement: Challenges, benchmarks, and solutions under natural light. IEEE Trans. Circ. Syst. Vid. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  56. Xi, D.; Qin, Y.; Wang, S. YDRSNet: An integrated Yolov5-Deeplabv3+ real-time segmentation network for gear pitting measurement. J. Intell. Manuf. 2021, 1–15. [Google Scholar] [CrossRef]
  57. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–120901. [Google Scholar]
Figure 1. The detailed structure diagram of MulTNet.
Figure 1. The detailed structure diagram of MulTNet.
Sensors 22 07224 g001
Figure 2. Examples of the marine animal segmentation obtained by FCN-8s, U-Net, Attention U-Net, RefineNet, ResUNet, Deeplabv3+, PSPNet, TrSeg, CKDNet, SegFormer, FAT-Net, and the proposed MulTNet.
Figure 2. Examples of the marine animal segmentation obtained by FCN-8s, U-Net, Attention U-Net, RefineNet, ResUNet, Deeplabv3+, PSPNet, TrSeg, CKDNet, SegFormer, FAT-Net, and the proposed MulTNet.
Sensors 22 07224 g002
Figure 3. Examples of different levels of blur obtained by the proposed MulTNet.
Figure 3. Examples of different levels of blur obtained by the proposed MulTNet.
Sensors 22 07224 g003
Figure 4. Examples of the marine animal segmentation obtained by FCN-8s, U-Net, Attention U-Net, RefineNet, ResUNet, Deeplabv3+, PSPNet, TrSeg, SegFormer, CKDNet, FAT-Net, and the proposed MulTNet.
Figure 4. Examples of the marine animal segmentation obtained by FCN-8s, U-Net, Attention U-Net, RefineNet, ResUNet, Deeplabv3+, PSPNet, TrSeg, SegFormer, CKDNet, FAT-Net, and the proposed MulTNet.
Sensors 22 07224 g004
Figure 5. The training losses of various segmentation networks for two datasets: (a) dataset of marine animals; (b) dataset of ISIC 2018.
Figure 5. The training losses of various segmentation networks for two datasets: (a) dataset of marine animals; (b) dataset of ISIC 2018.
Sensors 22 07224 g005
Table 1. A summary of the segmentation method mentioned in the literature.
Table 1. A summary of the segmentation method mentioned in the literature.
MethodologyYearHighlightsLimitations
FCN [4]: Fully convolutional network2015It is the first time CNN is used to extract features in the field of semantic segmentation.This model has poor robustness on the image detail and does not consider the relationship between pixels.
U-Net [6]: U-shaped end-to-end network with skip connection.2015This model exploits skip connections between encoder and decoder to decrease the loss of context information.This model generally shows poor performance in capturing details and explicitly building long-range dependency.
Attention U-Net [8]: Extension of standard U-Net model with attention mechanism.2018Suppress irrelevant areas in the input image and highlight the salient features of specific local areas.Attention gates (AGS) mainly focus on extracting the spatial information of the region of interest, and the ability to extract less relevant local regions is poor.
RefineNet [9]: a generic multi-path refinement network.2016It effectively exploits features in the downsampling process to achieve high-resolution prediction using long-distance residual connectivity.This model occupies large computing resources, resulting in low training speed and, secondarily, requires pre-trained weights on its backbones.
ResUNet [10]: Extension of standard U-Net model with residual units.2017Residual units simplify the training process of deep networks and skip connections could facilitate information propagation.This model cannot establish the dependency between pixels, which shows poor segmentation performance in blurred images.
Deeplabv3+ [16]: A segmentation network composed of atrous spatial pyramid pooling and decoder modules.2018This model can capture sharper object boundaries and extract multi-scale contextual information. Moreover, the resolution of the feature map of the encoder module can be arbitrarily controlled by atrous convolution.It adopts atrous convolution, which results in the loss of spatially continuous information.
PSPNet [17]: a pyramid scene parsing network embedding context features with PPM.2017It can aggregate the context information of different regions to improve the global information extraction abilityIt takes a long time to train and performs relatively poorly in detail handling.
CKDNet [18]: a cascade knowledge diffusion network composed of coarse-level segmentation, classification, and fine-level segmentation.2020This model adopts knowledge transfer and diffusion strategies to aggregate semantic information from different tasks to boost segmentation performance.This model consumes a lot of computing resources, resulting in low training speed.
U-Net-FS [19]: A optimal U-Net for free surface segmentation2020The outcomes of experiments show U-Net-FS can capture numerous types of free surfaces with a high dice accuracy of over 0.94.This model can only focus on local information and be trained with a single scale, so it cannot handle the change of image size well.
The fish image segmentation method is combined with the K-means clustering segmentation algorithm and mathematical morphology [20].2013The traditional K-means algorithm is improved by using the optimal number of clusters based on the peak number of the image gray histogram.It is sensitive to the selection of the initial value K and outliers, which will lead to poor segmentation performance.
Transunet [43]: Extension of standard U-Net model with transformer.2021This model can extract abundant global context by converting image features into sequences and exploiting the low-level CNN spatial information via a u-shaped architectural design.This model exploits transposed convolutional layers to restore feature maps. However, it usually results in the chess-board effect, which shows discontinuous predictions among adjacent pixels.
Swin transformer [49]: a hierarchical vision transformer with shifted windows2021The shift window scheme improves efficiency via limiting self-attention computation to non-overlapping local windows. Experimental results show SOTA performance in various tasks, including classification, detection, and segmentation tasks.Constraint by shift window operation. It is time-consuming that this model must be modified and retrained according to different sizes of input.
CoTr [50]: A hybrid network which combines CNN and transformer for 3D medical image segmentation.2021It exploits a deformable self-attention mechanism to decrease the spatial and computational complexities of building long-range relationships on multi-scale feature maps.Since both transformer and 3D volumetric data require a large amount of GPU memory, this method divides the data into small patches and deals with them one at a time, which causes the loss of features from other patches.
TrSeg [51]: A transformer-based segmentation architecture for multi-scale feature extraction.2021Different from existing networks for multi-scale feature extraction, this model incorporates a transformer to generate dependencies on original context information, which can adaptively extract multi-scale information well.This model has relatively few limitations on low-level semantic information extraction.
FAT-Net [52]: Feature adaptive transformer network for skin lesion segmentation.2021This model exploits both CNN and transformer encoder branches to extract rich local features and capture the important global context information.This segmentation network still has limitations when the color change of the image is too complex, and the contrast of the image is too low.
Table 2. Hyper-parameters of all the models used in this study.
Table 2. Hyper-parameters of all the models used in this study.
ParameterConfiguration
OptimizerSGD
Learning rate0.03
Weight decay0.01
Momentum0.9
Batch size6
Image size256*256
Activation functionGELU
Dropout0.3
Table 3. Comparison of various segmentation networks for the marine animal dataset.
Table 3. Comparison of various segmentation networks for the marine animal dataset.
ModelMPAIOU
(Sea Urchin)
IOU
(Sea Cucumber)
IOU
(Starfish)
MIOUAcc
FCN-8s36.87%27.41%1.68%15.99%35.48%97.05%
U-Net40.91%29.14%3.45%26.47%39.01%96.97%
Attention U-Net41.54%30.87%3.87%27.28%39.76%96.95%
RefineNet43.14%32.63%4.02%31.32%41.21%96.89%
ResUnet43.79%33.02%4.56%33.29%41.96%96.93%
DeepLabv3+44.24%34.21%5.22%33.23%42.40%96.96%
CKDNet46.05%39.26%4.31%33.58%43.54%97.09%
PSPNet45.38%37.67%5.62%34.13%43.62%97.07%
TrSeg46.32%40.67%4.71%33.47%43.98%97.11%
SegFormer [50]46.67%39.73%5.91%33.51%44.06%97.10%
FAT-Net47.17%42.97%5.06%34.35%44.87%97.16%
MulTNet47.69%44.14%5.21%35.93%45.63%97.27%
Table 4. Comparison of various segmentation networks for the public ISIC 2018 dataset.
Table 4. Comparison of various segmentation networks for the public ISIC 2018 dataset.
ModelMPAIOUMIOUAcc
FCN-8s89.98%67.47%79.03%92.12%
U-Net92.19%76.05%84.59%94.36%
Attention U-Net93.05%76.35%84.76%94.41%
RefineNet92.85%76.64%84.83%94.32%
ResUnet93.17%77.27%85.37%94.64%
DeepLabv3+93.31%78.21%85.99%94.92%
PSPNet93.08%79.52%86.77%95.15%
CKDNet93.13%79.90%86.99%95.21%
SegFormer93.40%80.19%87.21%95.32%
TrSeg93.22%80.37%87.31%95.33%
FAT-Net93.65%81.09%87.80%95.56%
MulTNet94.04%81.48%88.09%95.71%
Table 5. Comparison of computational costs.
Table 5. Comparison of computational costs.
ModelTotal ParamsTraining Speed (s/Iteration)Testing Speed (s/Image)
FCN-8s134 M0.7250.291
RefineNet85 M0.3810.148
CKDNet52 M0.3420.135
PSPNet71 M0.3170.124
Segformer64 M0.3120.127
TrSeg74 M0.2930.113
ResUnet67 M0.2850.110
DeepLabv3+55 M0.2380.096
AttentionU-Net42 M0.2030.092
FAT-Net30 M0.1940.089
U-Net32 M0.1670.075
MulTNet59 M0.1830.087
Table 6. Ablation experiments.
Table 6. Ablation experiments.
ISIC DatasetMarine Animal Dataset
ModelMIOUMPAAccMIOUMPAAcc
Transformer73.89%89.80%90.34%23.99%25.00%95.97%
DRCM-Transformer85.36%92.29%94.53%40.31%42.33%96.97%
MulTNet88.09%94.04%95.71%45.63%47.69%97.27%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, X.; Qin, Y.; Xi, D.; Ming, R.; Xia, J. MulTNet: A Multi-Scale Transformer Network for Marine Image Segmentation toward Fishing. Sensors 2022, 22, 7224. https://doi.org/10.3390/s22197224

AMA Style

Xu X, Qin Y, Xi D, Ming R, Xia J. MulTNet: A Multi-Scale Transformer Network for Marine Image Segmentation toward Fishing. Sensors. 2022; 22(19):7224. https://doi.org/10.3390/s22197224

Chicago/Turabian Style

Xu, Xi, Yi Qin, Dejun Xi, Ruotong Ming, and Jie Xia. 2022. "MulTNet: A Multi-Scale Transformer Network for Marine Image Segmentation toward Fishing" Sensors 22, no. 19: 7224. https://doi.org/10.3390/s22197224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop