Next Article in Journal
The Effect of (Mg, Zn)12Ce Phase Content on the Microstructure and the Mechanical Properties of Mg–Zn–Ce–Zr Alloy
Previous Article in Journal
Effect of Different Laser Groove Texture Collation Frequency on Tribological Properties of 0Cr17Ni7Al Stainless Steel
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved U-Net Image Segmentation Method and Its Application for Metallic Grain Size Statistics

1
National Center for Materials Service Safety, University of Science and Technology Beijing, Beijing 100083, China
2
Innovation Group of Marine Engineering Materials and Corrosion Control, Southern Marine Science and Engineering Guangdong Laboratory, Zhuhai 519080, China
3
School of Automation, Nanjing University of Science and Technology, Nanjing 211103, China
4
School of Information, Beijing Wuzi University, Beijing 101149, China
*
Author to whom correspondence should be addressed.
Materials 2022, 15(13), 4417; https://doi.org/10.3390/ma15134417
Submission received: 10 May 2022 / Revised: 17 June 2022 / Accepted: 19 June 2022 / Published: 22 June 2022

Abstract

:
Grain size is one of the most important parameters for metallographic microstructure analysis, which can partly determine the material performance. The measurement of grain size is based on accurate image segmentation methods, which include traditional image processing methods and emerging machine-learning-based methods. Unfortunately, traditional image processing methods can hardly segment grains correctly from metallographic images with low contrast and blurry boundaries. Moreover, the proposed machine-learning-based methods need a large dataset to train the model and can hardly deal with the segmentation challenge of complex images with fuzzy boundaries and complex structure. In this paper, an improved U-Net model is proposed to automatically accomplish image segmentation of complex metallographic images with only a small training set. The experiments on metallographic images show the significant advantage of the method, especially for the metallographic images with low contrast, a fuzzy boundary and complex structure. Compared with other deep learning methods, the improved U-Net scored higher in ACC, MIoU, Precision, and F1 indexes, among which ACC was 0.97, MIoU was 0.752, Precision was 0.98, and F1 was 0.96. The grain size was calculated based on the segmentation according to the American Society for Testing Material (ASTM) standards, producing a satisfactory result.

1. Introduction

Image segmentation is the process of dividing an image into several regions with similar properties. The traditional segmentation methods mainly include threshold-based segmentation methods [1], region-based segmentation methods [2], edge-detection-based segmentation methods [3], graph-theory-based segmentation methods [4], and energy functional-based segmentation methods [5]. Threshold-based segmentation methods divide image pixels into several categories by setting different feature thresholds, including the histogram threshold segmentation method [6], the iterative determination method [7], the inter-class variance threshold method [8], and the dynamic threshold segmentation method [9]. The region-based segmentation method divides the image into different regions according to the similarity criterion. It mainly includes the watershed segmentation algorithm [10], the level set segmentation algorithm [11], and the clustering segmentation algorithm [12]. The edge-based segmentation method is fast and effective for edge detection, but it is only suitable for images with relatively low noise and is not very suitable for complex data. It mainly includes the Canny operator [13], the Sobel operator [14], the Kirsch operator [15], and the Log operator [16]. The basic principle of the segmentation method based on graph theory is to remove specific edges and divide the graph into several subgraphs to achieve segmentation, mainly including GraphCut [17] and GrabCut [18]. The segmentation method based on energy functional mainly refers to the active contour model [19] and its improved algorithm. The active contour model can obtain a closed and smooth edge curve, but it is difficult to cope with the change of the model’s topology.
With the development of machine learning, some image segmentation based on machine learning arose, which can be subdivided into several directions, including the feature-encoder-based method [20], the region proposal network (RPN [21]), recurrent neural networks (RNNs) [22], upsampling or deconvolution-based method, and DeepLab [23]. RPN is an abbreviation for the region proposal network method, which is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. VGGNet [24] and ResNet [25] are two important methods in the field of feature extraction. The core idea of the regional proposal is to detect the color space and similarity matrix and then classify and predict according to the detection results. Ren et al. [26] merged RPN and Fast R-CNN [27] into a single network that achieved state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. RNN [28] is a network with sequence data as the input; recursion in the evolution direction of the sequence and all nodes are connected in a chain. RNN can learn from sequence data over a long time and retain memories along with sequences. This ability makes it useful for many computer vision tasks, including semantic segmentation and data annotation. Le et al. [29] proposed a multiview recurrent neural network (MV-RNN) approach for 3D mesh segmentation. Segmentation methods based on upsampling or deconvolution mainly include FCN [30], U-Net [31], and SegNet [20]. FCN is a fully convolutional network for semantic segmentation proposed by Jonathan Long. It treats images at the pixel level to solve the semantic segmentation problem. U-Net is a semantic segmentation network based on FCN. It can be trained from very few images and outperforms the previous best networks in the ISBI (Cell Image Segmentation by Electron Microscopy) challenge. SegNet is a pixel-level classification network with an encoder network and a corresponding decoder network. Differently from U-Net, SegNet uses pooled indices between the encoder and decoder instead of tensors, which makes SegNet less effective for the segmentation of images with fewer boundary pixels. The Deeplab method combines deep convolutional neural networks (DCNNs [32]) with a probability graph model (DenseCRFs [33]), which can complete per-pixel classification and achieve good image segmentation results. At the same time, there are various other deep learning methods applied to image recognition. DeCost et al. [34] used computer vision methods to collect various microstructural data, develop quantitative microstructural descriptors, and define objective categories of microstructures. Valente et al. [35] proposed a comparative analysis of a multilayer perceptron and self-organizing mapping topology used to segment a microstructure from metallographic images. Huang et al. [36] used the pretrained CNN representation to study the relationship between the processing conditions and the microstructure obtained by descent and visualization techniques. Chen et al. [37] used Mask R-CNN to complete the learning and recognition of potential characteristics of the microstructure of aluminum alloy. The U-Net++ network [38] proposed by Zhou overcomes the two limitations of U-Net: the optimal depth is a priori unknown and imposes unnecessary restrictive fusion schemes. Zhou et al. evaluated U-Net++ using six different medical image segmentation datasets and achieved good results. ResU-Net [39] is a semantic segmentation neural network which combines the strengths of residual learning, and U-Net is proposed for road area extraction. The network is built with residual units and has an architecture similar to that of U-Net. ResU-Net++ [40] is a fully automatic model for pixel-wise polyp segmentation improved from ResU-Net. It is an architecture created to address the need for more accurate segmentation of colorectal polyps found in colonoscopy examinations. A-DenseUNet [41] is an end-to-end image segmentation architecture. It aggregates multiscale semantic information to generate global features and encodes these features together with skip connection features to the decoder side. The model is able to learn multiscale semantic features at each level.
Although the segmentation methods above have been frequently applied to metallographic images, they have not worked well for complex metallographic images. One reason is that machine learning methods need a large number of labeled training images. Another fact is that in complex images such as metallographic images, the background of the segmentation task accounts for a large proportion, but the pixel proportion of the grain boundary that needs to be segmented is very small. Therefore, for such data sets with extremely uneven pixel distribution, it is difficult for the existing loss function to achieve the effect of network optimization, and this poor effect leads to a large deviation in the final measured grain size. Various properties of metals are closely related to the average size of their grains, and a deviation in grain size measurement will cause the predicted metal properties to be seriously inconsistent with the actual metal properties [42]. It is necessary to explore a new approach.
In this paper, an improved U-Net method is proposed for image segmentation, which can deal with complex images with a lower contrast and a blurred boundary. Moreover, the training of the improved U-Net model is simple and only needs a small dataset. Experimental observation shows that the single-phase metal image segmented by the method proposed in this study was much more accurate compared to those segmented by other methods. The potential of this method for quantitative analysis of microstructures was also investigated.
The main contributions of this paper are as follows. First of all, an improved U-Net model trained with a small training dataset was proposed, which worked well for segmentation of complex metallographic images and showed better performance than other models. Secondly, a loss function combining Focal loss and Dice loss was proposed to solve the problem of data imbalance in some datasets like metallographic images. Thirdly, DropBlock [43] and Batch Normalization [44] were used to prevent overfitting, and the Adam optimization algorithm [45] was used to optimize the improved U-Net model. Finally, based on the image segmentation results, the planimetric method and the intercept method were used to calculate the grain size according to the ASTM standard (E112-12), and satisfying results were obtained.
This paper is organized as follows. Section 2 describes the improved U-Net network structure, a specific loss function improvement, the Adam optimization algorithm, and the methods to prevent overfitting. In Section 3, the dataset, selection of evaluation metrics, details of network training, and comparison of experimental results are presented. The application of the method for an automatic rating of grain size also is described. Finally, Section 4 gives the conclusions of our work.

2. Improved U-Net

U-Net is a networking and training strategy that uses data augmentation to more efficiently utilize labeled data, which allows it to achieve good results with small sample data. However, the original U-Net model has several shortcomings in the loss function, optimization, and the overfitting problem. An improved U-Net network was proposed, which inherited the advantages of the U-Net model while improving in some other aspects. Firstly, Focal loss and Dice loss were combined as a new loss function of the improved U-Net model. It solved the problem of data imbalance in image segmentation and could achieve good segmentation results with a small training set of only 30 image samples. Secondly, to prevent the problem of overfitting caused by too little data, the DropBlock algorithm and Batch Normalization (BN) algorithm were used. Thirdly, Adam was chosen as the optimization algorithm, which could maximize the performance of the model and achieve the highest score in all evaluation metrics. The structure of the improved U-Net segmentation model is shown in Figure 1.

2.1. Improved Loss Function

The original U-Net uses binary cross-entropy [46] as the loss function. Binary cross-entropy is widely used in classification tasks and most semantic segmentation scenarios, but it also has obvious shortcomings. When the pixel distribution of the image is extremely unbalanced, the performance of the cross-entropy is very poor. When the image segmentation task only needs to segment the foreground and the background, the current number of foreground pixels is much smaller than the number of background pixels, that is to say, the number of y = 0 is much larger than the number of y = 1, which makes the model seriously biased towards the background, leading to poor results.
Lots of metallographic images have a heavy class imbalance (e.g., the boundary pixels only account for 3–5%). This imbalance results in a poor segmentation effect while the accuracy rate is very high. To solve this problem, a new loss function was designed.
Cross entropy examines every pixel in the image and then compares the class prediction (softmax or sigmoid) with the target vector. For dichotomies, the formula is as follows:
B C E = 1 N i = 1 N y i log p i + 1 y i log 1 p i
where N is the number of samples, y i represents the true sample label, and p i is the predicted probability of the category. Focal loss can solve the problem that the ratio of positive and negative samples is complete unbalanced in image segmentation, whose formula is as follows:
F L = 1 N i = 1 N α y i 1 p i γ log p i + ( 1 α ) 1 y i p i γ log 1 p i
where α is the weight factor, by adding ( 1 p i ) γ , the loss of the sample with a large prediction probability is reduced, while the loss of the sample with a small prediction probability is increased, so as to strengthen the attention on the positive sample and improve the phenomenon of target imbalance.
The Dice coefficient is derived from Lee Raymond Dice [47]. Dice is a set similarity measurement function, which can be used to calculate the similarity between two samples (value range is [0, 1]). The Dice coefficient can also be viewed as a loss function:
DC ( A , B ) = 2 TP 2 TP + FP + FN = 2 A B A + B
where A is the predicted value, and B is the truth value. The meanings of TP, FP, TN, and FN are shown in Table 1.
In this paper, after modifying the Dice coefficient, a suitable Dice loss function can be obtained:
DL = TP TP + α FN + β FP
Expand the description of Equation (4), then:
DL = i = 1 N p i g i p i g i + α g i ( 1 p i ) + β p i ( 1 g i )
where α and β are the trade-offs of penalties for false negatives and false positives, and g i is the ground truth. Combining the Dice loss and Focal loss, then:
Loss = λ FL + DL
Then expanding the description of this loss function:
Loss = i = 1 N p i g i p i g i + α g i ( 1 p i ) + β p i ( 1 g i ) λ 1 N i = 1 N α y i 1 p i γ log p i + ( 1 α ) 1 y i p i γ log 1 p i
where λ is the trade-off between Dice loss and Focal loss, and λ is set to be 0.5 based on the performance on the validation set. In some scenarios (such as metallographic images), most negative samples are located in the background area, and few negatives are located in the transition area between the foreground and the background. The samples located in the background area are relatively easy to be classified. Meanwhile, the corresponding score during training is large, and the loss is relatively small. When calculating the reverse gradient, the easy negative example has a limited impact on parameter convergence. As a result, hard negative samples are difficult to be updated and have poor results.
Considering the unbalanced pixels, more weight values are added on the boundary. By combining Focal Loss and Dice Loss and setting the parameters to α = 0.25 and β = 2 , a better effect than that of other loss functions is achieved.

2.2. Adam Optimization

The optimization problem is one of the core problems of machine learning. A suitable optimizer can improve efficiency and save computing resources, which can play a key role in improving the speed and effect of machine learning.
In this paper, after comparing different optimizers, Adam was used as the optimizer. Unlike a traditional stochastic gradient descent, which maintains a single learning rate, Adam designs independent adaptive learning rates for different parameters by computing the first- and second-moment estimates of the gradient. It combines the advantages of AdaGrad and RMSProp.
The AdaGrad optimization algorithm is called an adaptive learning rate optimization algorithm. Each time a batch size of data is used for a parameter update, the algorithm calculates the gradients of all the parameters and then accumulates the squared sum of the gradients of the parameters to a variable s with an initial value of 0. When updating this parameter, the learning rate will vary with s.
AdaGrad reserves a learning rate for each parameter to improve performance on sparse gradients, while RMSProp adaptively preserves the learning rate for each parameter based on the mean of the nearest magnitude of the weight gradient.
By combining the advantages of AdaGrad and RMSProp, the Adam optimization algorithm has excellent performance on both nonstationary and online problems. Specifically, the algorithm computes exponential moving averages of gradients, and the hyperparameters beta1 and beta2 control the decay rate of these moving averages. The initial value of the moving average and the beta1, beta2 values are close to 1 (recommended value), so the bias of the moment estimates is close to 0. The bias is boosted by first computing the biased estimate and then computing the bias-corrected estimate.

2.3. Prevention of Overfitting

When the training sample is small while the model is more complex, it often faces the problem of overfitting. Considering that our training data were few, three methods were used to prevent the overfitting problem.
First, more data was invested in training. Data augmentation methods such as panning, scaling, shading, and flipping images were used to augment the data. Specifically, random rotations were adopted in the range (0, 2π) (including horizontal and vertical mirror symmetry) and set the scaling factor between 0.8 and 1.2 for all training and label data. Second, the DropBlock method was used. DropBlock is a regularization method suitable for convolutional layers. The DropBlock method drops features in adjacent and related feature regions that become blocks. It can achieve the purpose of generating a simpler model and introduce the concept of learning partial network weights in each iteration training to compensate for the weight matrix, thereby reducing overfitting. Third, Batch Normalization was used for training. When training with Batch Normalization, you can see that one training example appears with other examples in the mini-batch, and the training network no longer produces deterministic values for a given training example. The same sample and different samples form a mini-batch, and their outputs are different. It means that the output of the same sample is related to both the sample itself and other samples that belong to the same mini-batch as this sample. This correlation alleviates the overfitting problem of the model.

3. Results and Discussion

In this study, the improved U-Net was adopted to segment images at a high level of abstraction. The specific work steps are shown in Figure 2. To train a better improved U-Net network, after labeling the image, the given training set was preprocessed, including image size adjustment, Gaussian smoothing, image denoising, contrast enhancement, and data augmentation. After the original image was segmented by the improved U-Net model and other segmentation methods, the segmentation results were compared and analyzed. Firstly, the segmentation effects of different segmentation methods were compared. Secondly, the performance of different loss functions was compared. Thirdly, the grain size of the image segmented by the improved U-Net network was measured, and the measurement error was calculated.

3.1. Dataset

In this paper, many typical single grain metallographic images were collected as our experiment dataset, called SGMD. The dataset contained 60 image samples, including 690 alloy, nickel base alloy, low-carbon steel alloy, austenitic steel, pure iron, yellow brass, and ferrite steel phase metallographic pictures.
The dataset was divided into the training set, the validation set, and the test set. The training set contained 30 image samples, the validation set contained 15 image samples, and the test set contained 15 image samples. The dataset was randomly divided one hundred times, and experiments were executed separately. The averaged performance value was used for comparison.

3.1.1. Image Labeling

By using Labelme (https://github.com/wkentaro/labelme, accessed on 11 May 2020), our metallographic images were annotated, and their boundaries were obtained. Figure 3 shows the process of image labeling. First, Labelme was used to label Figure 3a (original image) to generate Figure 3b. Because the red border of the label was too thick, the small crystals were almost covered, so it was necessary to refine the image that was labeled for the first time. After binarization, the skeleton was extracted by the morphological method, and the skeleton was retained as a single pixel, as shown in Figure 3c. After being processed by the skeleton extraction algorithm, the border was only one pixel. The proportion of border pixels was very small in the background, and the training of the model would have been very difficult. The boundaries were thickened with the method of erosion in morphology so that it could be trained normally, and the final result is shown in Figure 3d.

3.1.2. Image Preprocessing

During image preprocessing, median filtering and Gaussian filtering were used for simple denoising. Median filtering is nonlinear filtering that preserves edge information by removing salt and pepper noise, while Gaussian filtering is smooth linear filtering that can effectively remove Gaussian noise.
Since the grains in the metallographic images were very dense and irregular in size, it was difficult to manually mark them, so the data were relatively small. To get good training results with less data, data augmentation was the most effective way. This paper mainly used elastic deformation, translation, rotation, light and dark changes, and other methods to expand the data. With sufficient data in this way, the training effect of the network model could be improved, and the network model had invariance and good robustness. The preprocessed images are shown in Figure 4.

3.2. Evaluation Metrics

To accurately measure the performance of segmentation models, several metrics were adopted, including ACC, Precision, Recall, F1-score, the Dice coefficient, and Mean IoU (MIoU). They are widely used in image segmentation, defined as follows.
ACC [48] is the simplest metric, which is the percentage of pixels that are correctly marked relative to the total pixels (Equation (6)).
ACC = TP + TN TP + TN + FP + FN
Precision (also called positive predictive value) is the fraction of relevant instances among the retrieved instances, which shows the correct detection ability of the response model for abnormal data. It is calculated as Equation (7):
Precision = TP TP + FP
Recall refers to the ratio of the total number of correctly identified abnormal samples to the number of true abnormal samples. It reflects the ability of the response model to identify abnormal data, calculated by Equation (8):
Recall = TP TP + FN
F1-score is a combination of precision and recall. The robustness of the classification model is positively correlated with the F1 score. It is calculated as Equation (9).
F 1 = 2 TP 2 TP + FP + FN = 2 × Precision × Recall Precision + Recall
The Dice coefficient is the most commonly used evaluation index in semantic segmentation. The higher the Dice coefficient, the higher the similarity between two samples. The Dice score is defined as Equation (10).
DC ( A , B ) = 2 TP 2 TP + FP + FN = 2 A B A + B
where A is the predicted segmentation and B is the ground truth. Intersection over Union (IoU) is another metric used to evaluate the predictions from an image segmentation model. The IoU is defined as Equation (11).
IoU ( A , B ) = TP TP + FP + FN = A B A B
Mean-IoU is a metric that takes the IoU over all of the classes and takes the mean of them, calculated as Equation (12).
MIoU = 1 k + 1 i = 0 k TP FN + FP + TP

3.3. Model Training

This part mainly introduces more details regarding the implementation environment, data augmentation strategies, and hyperparameters when training the model.
The model was implemented by the Keras package [49]. In the process of training with the basic U-Net, batch normalization, dropout regularization, weight decay regularization, and data augmentation were combined. To reduce the global brightness difference between different samples and data sets, local histogram equalization was performed on the trained images. The number of channels of each convolution layer was increased by 0.5 times, and the image input dimension was set to 512 × 512 pixels, which improved the fitting and generalization ability of the model.
For data augmentation, the data were randomly rotated in the range (0, 2π), including horizontal and vertical mirror symmetry, and scaled between 0.8 and 1.2 for all training data and labeled images to alleviate the overfitting problem.
Correctly tuning the values of hyperparameters is the key to obtaining good performance of machine learning models. According to our experiences, the initial learning rate was set to 0.015, and decay was set according to Equation (13). After much repeated training, the loss no longer fluctuated on both the validation set and the training set after 40 epochs. Consequently, the epoch was set to 50, and batch size was set to 16. The improved training network loss could reach 0.0112, as shown in Figure 5.
lr = lr × ( 1 iterations total _ iterations ) 0.9
In addition, to achieve good performance on the validation set, adaptive learning rate adjustment and early stopping criteria were also set. The specific rules were as follows.
If there was no improvement in validation loss over five epochs, the learning rate began to decline.
If the validation loss did not improve over the next ten epochs, the training was terminated.
After multiple sets of experimental comparisons, the Focal loss and Dice loss were finally selected as the loss function. When the parameters were set to α = 0.25, β = 2, the best effect could be obtained, and MIoU could reach 0.88. Since the adjustment factor reduces the contribution of simple sample loss, reduces the rate of simple sample weighting, and expands the acceptance range of low loss samples, it is more beneficial for the study of difficult samples.
Table 2 shows the performance of different optimizers. Adam performed the best of all the metrics, so it was chosen as the optimization algorithm.

3.4. Performance Comparison with Traditional Image Segmentation Methods

To verify the performance of the proposed method, our method was compared with an advanced material image processing software called Mipar and several representative image segmentation methods, including the Morphological method, the Canny operator, and the Watershed method.

3.4.1. Morphological Method

The Morphological method utilizes morphological operations, such as erosion, dilation, opening, closing, and top-hat transformation, to extract, modify, and manipulate the features presented in the image based on their shapes [50]. In order to facilitate the extraction of grain boundaries and quantitative information, the grayscale image is converted into a binary image. After binarization, there is a lot of point-like salt and pepper noise, which is removed by the median filtering method. In order to preserve the linear boundary, a similar linear grain substructure adopts the dot net division method and uses eight connected neighborhood pixel models to search the image edges.

3.4.2. Canny Operator

Canny operator is a strong edge detection operator proposed by Canny in 1986. It is an optimal approximation operator for the product of SNR and location. Because the Canny operator adopts the hysteresis threshold method, the false edges generated by noise and grain internal texture can be suppressed well, and it has a good effect on the edge detection of the image of the metallographic structure.

3.4.3. Watershed Method

The Watershed algorithm is essentially an image segmentation method based on topological theory. It mainly divides images according to their domain characteristics, but for metallographic images with too much noise, it is easy to produce oversegmentation.
In this metallurgical micrograph image segmentation experiment, the U-Net, U-Net++, and improved U-Net were compared with the traditional segmentation algorithms and typical software (Mapir, San Diego, CA, USA). The specific effect is shown in Figure 6. The result of the Morphological method is shown in Figure 6b. Because the noise was mistaken for time boundary pixels, the boundary of the segmentation result was too thick. The method based on the Canny operator in Figure 6c could clearly detect the boundary, but it was difficult to remove noise, which would eventually affect the final grain size statistics. The Watershed-based image segmentation algorithm in Figure 6d was prone to oversegmentation for images with more noise, so it could not be used in the dataset. Figure 6e is the segmentation result of advanced material image processing software called Mipar (http://keple.cn/170195-170195.html, accessed on 6 June 2021), but the result was very poor. The original U-Net model was very poorly realized in this experiment. As seen from Figure 6f, the segmentation was very fuzzy and the boundary could be recognized basically. The segmentation result of U-Net++ is shown in Figure 6g. Compared with that of U-Net, it had better performance, but it did not make much progress in segmenting complex regions. Figure 6h shows the result of the improved U-Net model, whose segmentation effect was the best, with the noise being completely filtered out. Thus, the improved U-Net was adopted as the segmentation method.

3.5. Performance Comparison with Other Deep Learning Methods

In this experiment, our method was compared with several deep learning methods suitable for microscopic image segmentation, including FCN, SegNet, DeepLabV3, Mask R-CNN, U-Net, ResU-Net, A-DenseU-Net, ResU-Net++, and U-Net++. The experimental results are shown in Table 3. Compared with those of other deep learning methods, the improved U-Net achieved good results in all indicators. The improved U-Net scored higher in ACC, MIoU, Precision, and F1 indexes, among which ACC was 0.97, MIoU was 0.752, Precision was 0.98, and F1 was 0.96. The improved U-Net of the proposed loss function could guide the practice to pay more attention to the direction of the boundary pixels. To prevent overfitting, the regularization terms are introduced. In this paper, Adam tuning is chose based on the optimal learning rate and batch size. According to the experiment, Adam tuning have a good performance than other optimizers.
Since the convergence effect is important for a deep learning method, the convergence effects of these networks were compared, and the results are shown in Figure 7. The loss began to converge after twenty epochs. The improved U-Net, U-Net++, ResU-Net, A-DenseU-Net, ResU-Net++, U-Net, Mask R-CNN, and DenseNet converged faster, while FCN, SegNet, and Deeplab V3 converged relatively slowly. After forty epoch trainings, the training loss was stable in a converging state. As a whole, the improved U-Net had a faster convergence speed and lower loss value, showing better performance.

3.6. Performance Comparison with Different Loss Functions

To solve the problem of extremely unbalanced image pixels, many improved loss functions have been proposed, and the effects of these loss functions on the symmetry of metallographic images have been studied. At present, few papers have done this work in the field of metallographic images. In this paper, five different loss functions were applied to the segmentation of metallographic images. The loss functions used in this paper include Cross entropy loss, Dice loss, Tversky loss, Tversky and Focal generalized loss, and Focal loss.
After the comparison of several groups of experiments, the best result of each loss function was selected for comparison, and the combination of Focal loss and Dice loss was finally selected as the loss function. When the parameter was set to α = 0.25, β = 2, the overall effect was the best, and the MIoU could reach 0.88. The adjustment factor reduces the impact of the loss of simple samples, reduces the weighting ratio of simple samples, and expands the acceptance range of low loss samples, so it is more favorable for the study of difficult samples. The final experimental results are shown in Table 4.
It can be seen that ACC of the six methods could reach more than 95%. Since the ACC evaluation indices are very similar, it is difficult to evaluate the performance of loss functions.
Compared with ACC, precision is more suitable for evaluating the performance of the metallographic image segmentation method because it avoids the interference of the large background. The experimental results show that the frame of this paper was reliable.
The experimental results also show that the loss function had a very important influence on the image segmentation effect. Among them, Focal loss and Dice loss were combined, and its parameter was set to 0.25 after many explorations, and the effect was best when it was set to 2. Among them, Loss, ACC, Precision, Recall, MIoU, Dice, and F1 were 0.102, 0.987, 0.931, 0.965, 0.88, 0.761, and 0.958, respectively. All the scores were relatively high, and the best results were achieved.

3.7. The Application of Grain Size Rating

The quantitative analysis of grain size is of great theoretical and practical significance to the study of materials. The American Society for Testing Material (ASTM) standard E112-12 recommends the planimetric and intercept methods for more accurate grain size measurements. The grain size G can be calculated as follows:
G = 3.321928 log 10 N A 2.954
G = 6.643856 log 10 l ¯ 3.288
The planimetric method is shown in Equation (16), and the intercept method is shown in Equation (17). N A represents the total number of grains per unit area, and l ¯ represents the average of multiple cross-sections. The result of the planimetric method and intercept method is shown in Figure 8a and Figure 8b, respectively. It can be seen that the planimetric method and intercept method could produce the same level of precision.
In this paper, the planimetric method was used to calculate the grain size. To measure the grain area step:
  • Select a rectangular area and calculate its area, as shown in Figure 8a.
  • Calculate the number of grains: N = N + 1 2 N ( N is the number of grains in the box, N is the number of grains across the box boundary, shown in Table 5). The result calculated by the formula method differs very little from the result manually calculated, and a high accuracy rate is achieved.
  • The grain area is obtained by dividing the rectangular area by the number of grains.
  • Calculate the number of grains in the unit area (mm2) and the grain size level, shown in Table 6.
Equation (16) gives the 95% confidence interval (95% CI) [51], and Equation (17) gives the relative accuracy percentage (%RA) [52], which can describe the accuracy of an automatic measurement.
95 % C I = ± t s n
% R A = 95 % C I l
As shown in Table 6, the difference between the numbers of grain measured manually and automatically was kept within 2. The accuracy was very high, and the error rate was no higher than 0.01. It can also be seen that the values of 95% CI and %RA were both less than 10. Therefore, the measured results were acceptable.

4. Conclusions and Future Work

An improved U-Net model with less training data was proposed to segment images. The model only trained with sixty images could produce satisfactory segmentation results, especially for fuzzy boundary metallographic images with a complex microstructure. The proposed method was compared with different image segmentation methods, including the Morphological method, the Canny operator, the Watershed method, FCN, SegNet, DeepLabV3, DenseNet, Mask R-CNN, U-Net, ResU-Net, A-DenseU-Net, ResU-Net++, and U-Net++. The comparison results showed that the improved U-Net was superior to the traditional metallographic image processing methods and other machine learning-based methods. It had higher scores on ACC, MIoU, Precision, and F1, with ACC of 0.97, MIoU of 0.752, Precision of 0.98, and F1 of 0.96.

Author Contributions

Writing—original draft preparation, P.S.; writing—review and editing, software, and validation, L.J.; software, M.D.; validation, L.Y., W.F. and L.D. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was funded by the National Natural Science Foundation of China (No. 92060103), the Joint Fund of Ministry of Education of China for Equipment Pre-research (No. 641A0202022260), the Construction Project of Innovation Group of Southern Marine Science and Engineering Guangdong Laboratory (Zhuhai) (No. 311020012), and the Research Program of Beijing Municipal Education Commission (No. KM201910037001 and No. KM201910037186).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Access to download the code and data can be found on Dropbox: https://bit.ly/3zOaKtX; accessed on 18 June 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bhargavi, K.; Jyothi, S. A survey on threshold based segmentation technique in image processing. Int. J. Innov. Res. Dev. 2014, 3, 234–239. [Google Scholar]
  2. Kaganami, H.G.; Beiji, Z. Region-based segmentation versus edge detection. In Proceedings of the 2009 Fifth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, Kyoto, Japan, 12–14 September 2009; pp. 1217–1221. [Google Scholar]
  3. Huang, Y.-C.; Tung, Y.-S.; Chen, J.-C.; Wang, S.-W.; Wu, J.-L. An adaptive edge detection based colorization algorithm and its applications. In Proceedings of the 13th Annual ACM International Conference on Multimedia, Singapore, 6–11 November 2005; pp. 351–354. [Google Scholar]
  4. Parihar, V.R.; Thakur, N.V. Graph theory based approach for image segmentation using wavelet transform. Int. J. Image Process. 2014, 8, 255–277. [Google Scholar]
  5. Bresson, X.; Vandergheynst, P.; Thiran, J.-P. A priori information in image segmentation: Energy functional based on shape statistical model and image information. In Proceedings of the 2003 International Conference on Image Processing (Cat. No. 03CH37429), Catalonia, Spain, 14–18 September 2003; p. III-425. [Google Scholar]
  6. Raju, P.D.R.; Neelima, G. Image segmentation by using histogram thresholding. Int. J. Comput. Sci. Eng. Technol. 2012, 2, 776–779. [Google Scholar]
  7. Rezaiee-Pajand, M.; Entezami, A.; Shariatmadar, H. An iterative order determination method for time-series modeling in structural health monitoring. Adv. Struct. Eng. 2018, 21, 300–314. [Google Scholar] [CrossRef]
  8. Arifin, A.Z.; Asano, A. Image segmentation by histogram thresholding using hierarchical cluster analysis. Pattern Recognit. Lett. 2006, 27, 1515–1521. [Google Scholar] [CrossRef]
  9. Yogarajah, P.; Condell, J.; Curran, K.; Cheddad, A.; McKevitt, P. A dynamic threshold approach for skin segmentation in color images. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, 26–29 September 2010; pp. 2225–2228. [Google Scholar]
  10. Li, C.; Min, J.; JianXun, C. Image segmentation using iterative watersheding plus ridge detection. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 4033–4036. [Google Scholar]
  11. Kandhasamy, J.P.; Balamurali, S.; Kadry, S.; Ramasamy, L.K. Diagnosis of diabetic retinopathy using multi level set segmentation algorithm with feature extraction using svm with selective features. Multimed. Tools Appl. 2020, 79, 10581–10596. [Google Scholar] [CrossRef]
  12. Salvador, S.; Chan, P. Determining the number of clusters/segments in hierarchical clustering/segmentation algorithms. In Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence, Boca Raton, FL, USA, 15–17 November 2004; pp. 576–584. [Google Scholar]
  13. Eisenbach, M.; Stricker, R.; Seichter, D.; Amende, K.; Gross, H.M. How to get pavement distress detection ready for deep learning? A systematic approach. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017. [Google Scholar]
  14. Gao, W.; Lei, Y.; Zhang, X.; Liu, H. An improved Sobel edge detection. In Proceedings of the IEEE International Conference on Computer Science & Information Technology, Chengdu, China, 9–11 July 2010. [Google Scholar]
  15. Decordier, I.; Papine, A.; Vande Loock, K.; Plas, G.; Soussaline, F.; Kirsch-Volders, M. Automated image analysis of micronuclei by IMSTAR for biomonitoring. Mutagenesis 2011, 26, 163–168. [Google Scholar] [CrossRef] [Green Version]
  16. Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1980, 207, 187–217. [Google Scholar]
  17. Kwatra, V.; Schödl, A.; Essa, I.; Turk, G.; Bobick, A. Graphcut textures: Image and video synthesis using graph cuts. ACM Trans. Graph. (TOG) 2003, 22, 277–286. [Google Scholar] [CrossRef]
  18. Rother, C.; Kolmogorov, V.; Blake, A. “GrabCut” interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004, 23, 309–314. [Google Scholar] [CrossRef]
  19. Jiang, H.; Cheng, Q. Automatic 3D segmentation of CT images based on active contour models. In Proceedings of the 2009 11th IEEE International Conference on Computer-Aided Design and Computer Graphics, Huangshan, China, 19–21 August 2009; pp. 540–543. [Google Scholar]
  20. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  21. Ning, J.; Zhang, L.; Zhang, D.; Wu, C. Interactive image segmentation by maximal similarity based region merging. Pattern Recognit. 2010, 43, 445–456. [Google Scholar] [CrossRef]
  22. Graves, A.; Mohamed, A.-r.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  23. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Wang, Z.; Zheng, X.; Li, D.; Zhang, H.; Yang, Y.; Pan, H. A VGGNet-like approach for classifying and segmenting coal dust particles with overlapping regions. Comput. Ind. 2021, 132, 103506. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  27. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  28. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  29. Le, T.; Bui, G.; Duan, Y. A multi-view recurrent neural network for 3D mesh segmentation. Comput. Graph. 2017, 66, 103–112. [Google Scholar] [CrossRef]
  30. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  32. Rawat, W.; Wang, Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef]
  33. Wolf, D.; Prankl, J.; Vincze, M. Fast semantic segmentation of 3D point clouds using a dense CRF with learned parameters. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 4867–4873. [Google Scholar]
  34. DeCost, B.L.; Holm, E.A. A computer vision approach for automated analysis and classification of microstructural image data. Comput. Mater. Sci. 2015, 110, 126–133. [Google Scholar] [CrossRef] [Green Version]
  35. Valente, I.R.S.; Cortez, P.C.; Neto, E.C.; Soares, J.M.; de Albuquerque, V.H.C.; Tavares, J.M.R. Automatic 3D pulmonary nodule detection in CT images: A survey. Comput. Methods Progr. Biomed. 2016, 124, 91–107. [Google Scholar] [CrossRef] [Green Version]
  36. Huang, T.; Yang, G.; Tang, G. A fast two-dimensional median filtering algorithm. IEEE Trans. Acoust. Speech Signal Process. 1979, 27, 13–18. [Google Scholar] [CrossRef] [Green Version]
  37. Chen, D.; Guo, D.; Liu, S.; Liu, F. Microstructure instance segmentation from aluminum alloy metallographic image using different loss functions. Symmetry 2020, 12, 639. [Google Scholar] [CrossRef] [Green Version]
  38. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef] [Green Version]
  39. Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef] [Green Version]
  40. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H.D. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  41. Safarov, S.; Whangbo, T.K. A-DenseUNet: Adaptive densely connected UNet for polyp segmentation in colonoscopy images with atrous convolution. Sensors 2021, 21, 1441. [Google Scholar] [CrossRef]
  42. Peregrina-Barreto, H.; Terol-Villalobos, I.; Rangel-Magdaleno, J.; Herrera-Navarro, A.; Morales-Hernández, L.; Manríquez-Guerrero, F. Automatic grain size determination in microstructures using image processing. Measurement 2013, 46, 249–258. [Google Scholar] [CrossRef]
  43. Ghiasi, G.; Lin, T.-Y.; Le, Q.V. Dropblock: A regularization method for convolutional networks. arXiv 2018, arXiv:1810.12890. [Google Scholar]
  44. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  45. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  46. Li, C.H.; Lee, C. Minimum cross entropy thresholding. Pattern Recognit. 1993, 26, 617–625. [Google Scholar] [CrossRef]
  47. Zou, K.H.; Warfield, S.K.; Bharatha, A.; Tempany, C.M.; Kaus, M.R.; Haker, S.J.; Wells, W.M., III; Jolesz, F.A.; Kikinis, R. Statistical validation of image segmentation quality based on a spatial overlap index1: Scientific reports. Acad. Radiol. 2004, 11, 178–189. [Google Scholar] [CrossRef] [Green Version]
  48. Wang, Z.; Yoon, S.; Xie, S.J.; Lu, Y.; Park, D.S. A high accuracy pedestrian detection system combining a cascade AdaBoost detector and random vector functional-link net. Sci. World J. 2014, 2014, 105089. [Google Scholar] [CrossRef] [PubMed]
  49. Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems; O’Reilly Media, Inc.: Newton, MA, USA, 2019. [Google Scholar]
  50. Mukhopadhyay, S.; Chanda, B. Multiscale morphological segmentation of gray-scale images. IEEE Trans. Image Process. 2003, 12, 533–549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Paredes-Orta, C.A.; Mendiola-Santibañez, J.D.; Manriquez-Guerrero, F.; Terol-Villalobos, I.R. Method for grain size determination in carbon steels based on the ultimate opening. Measurement 2019, 133, 193–207. [Google Scholar] [CrossRef]
  52. Li, M.; Chen, D.; Liu, S.; Liu, F. Grain boundary detection and second phase segmentation based on multi-task learning and generative adversarial network. Measurement 2020, 162, 107857. [Google Scholar] [CrossRef]
Figure 1. The structure of the improved U-Net segmentation model.
Figure 1. The structure of the improved U-Net segmentation model.
Materials 15 04417 g001
Figure 2. Specific work steps’ flow chart.
Figure 2. Specific work steps’ flow chart.
Materials 15 04417 g002
Figure 3. Process of image annotation. (a) Original image; (b) annotated image; (c) single-pixel skeleton image; (d) bolded image.
Figure 3. Process of image annotation. (a) Original image; (b) annotated image; (c) single-pixel skeleton image; (d) bolded image.
Materials 15 04417 g003
Figure 4. Examples of metallographic images. (ad) Four metallographic images and (eh) preprocessed images.
Figure 4. Examples of metallographic images. (ad) Four metallographic images and (eh) preprocessed images.
Materials 15 04417 g004
Figure 5. The training and validation loss of the model’s training.
Figure 5. The training and validation loss of the model’s training.
Materials 15 04417 g005
Figure 6. The segmentation result of different segmentation methods. (a) The original image; (b) Morphological methods; (c) Canny algorithm; (d) Watershed algorithms; (e) Mipar; (f) U-Net; (g) U-Net++; (h) improved U-Net.
Figure 6. The segmentation result of different segmentation methods. (a) The original image; (b) Morphological methods; (c) Canny algorithm; (d) Watershed algorithms; (e) Mipar; (f) U-Net; (g) U-Net++; (h) improved U-Net.
Materials 15 04417 g006
Figure 7. Training losses of different segmentation methods.
Figure 7. Training losses of different segmentation methods.
Materials 15 04417 g007
Figure 8. Grain size measurement results according to ASTM E112-12. (a) The planimetric method; (b) the intercept method.
Figure 8. Grain size measurement results according to ASTM E112-12. (a) The planimetric method; (b) the intercept method.
Materials 15 04417 g008
Table 1. The meanings of TP, FP, TN, and FN.
Table 1. The meanings of TP, FP, TN, and FN.
TerminologyFull NameMeaning
TPTrue positiveThe number of positive examples predicted correctly.
FPFalse positiveThe number of positive example prediction errors.
TNTrue negativeThe number of negative examples predicted correctly.
FNFalse negativeThe number of negative example prediction errors.
Table 2. Overall performance for different optimizers.
Table 2. Overall performance for different optimizers.
OptimizerLearning RateACCDiceMIoUPrecision
Adam0.0150.9650.8780.8940.981
SGD0.0150.9410.7830.8510.950
Adadelta0.0150.9120.7900.7950.891
AdaGrad0.0150.9340.8170.6100.887
Table 3. The metrics’ results of different segmentation methods.
Table 3. The metrics’ results of different segmentation methods.
MethodsACCDiceMIoUPrecisionRecallF1
FCN0.6540.6780.3410.850.8760.86
SegNet0.6790.7830.4510.950.8530.898
Deeplab V30.910.7900.5950.7340.9830.840
DenseNet0.890.9170.6100.6570.9420.774
Mask R-CNN0.870.9370.5560.8940.8710.882
U-Net0.890.9100.7000.9040.8800.892
ResU-Net0.9050.9120.6210.8640.880.872
A-DenseU-Net0.8430.9160.6840.920.8550.886
ResU-Net++0.9190.9270.7410.890.8030.844
U-Net++0.960.9150.7180.9570.890.922
Improved U-Net0.970.930.7520.980.9400.960
Table 4. The performance of different loss functions on our model.
Table 4. The performance of different loss functions on our model.
Loss FunctionLossAccPrecisionRecallMIoUDiceF1
Cross entropy0.1340.9670.6500.5580.3350.4830.600
Focal loss0.6310.9730.8590.1220.0130.4030.210
Dice loss0.6540.9520.8320.8070.3420.7860.819
Tversky0.1120.9670.7830.7830.6320.4310.783
Focal and Dice0.1020.9870.9510.9650.8800.7610.958
Tversky and focal
Generalized
0.1320.9680.6530.5730.6000.6630.610
Table 5. Typical metallographic grain number statistics in SGMD.
Table 5. Typical metallographic grain number statistics in SGMD.
Image IDHumanAutomatic CalculationError NumberError Rate
1173174+10.008
2909000
3110114+40.036
46968−10.014
520020000
6151152+10.007
7160159−10.006
817117000
930030000
10190188−20.1
Average 0.017
Table 6. Typical grain size level statistics in SGMD.
Table 6. Typical grain size level statistics in SGMD.
Image IDG (ASTM) ± 95 %   C I %RA
16.900.1635.41
25.700.2668.32
37.360.1515.02
47.890.2016.04
57.950.32110.1
67.900.1314.51
76.680.1635.41
86.730.2046.21
96.920.2106.30
107.500.1093.01
Average7.0530.19196.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shi, P.; Duan, M.; Yang, L.; Feng, W.; Ding, L.; Jiang, L. An Improved U-Net Image Segmentation Method and Its Application for Metallic Grain Size Statistics. Materials 2022, 15, 4417. https://doi.org/10.3390/ma15134417

AMA Style

Shi P, Duan M, Yang L, Feng W, Ding L, Jiang L. An Improved U-Net Image Segmentation Method and Its Application for Metallic Grain Size Statistics. Materials. 2022; 15(13):4417. https://doi.org/10.3390/ma15134417

Chicago/Turabian Style

Shi, Peng, Mengmeng Duan, Lifang Yang, Wei Feng, Lianhong Ding, and Liwu Jiang. 2022. "An Improved U-Net Image Segmentation Method and Its Application for Metallic Grain Size Statistics" Materials 15, no. 13: 4417. https://doi.org/10.3390/ma15134417

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop