Next Article in Journal
A Novel Low-Cost Compact High-Performance Flower-Shaped Radiator Design for Modern Smartphone Applications
Next Article in Special Issue
Core–Shell Particles: From Fabrication Methods to Diverse Manipulation Techniques
Previous Article in Journal
A Novel Multi-Modal Teleoperation of a Humanoid Assistive Robot with Real-Time Motion Mimic
Previous Article in Special Issue
The Effect of Non-Uniform Magnetic Field on the Efficiency of Mixing in Droplet-Based Microfluidics: A Numerical Investigation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Recognition Algorithm of Two-Phase Flow Patterns Based on GoogLeNet+5 Coord Attention

1
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
2
School of Mechanics and Engineering Science, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Micromachines 2023, 14(2), 462; https://doi.org/10.3390/mi14020462
Submission received: 31 December 2022 / Revised: 14 February 2023 / Accepted: 15 February 2023 / Published: 16 February 2023
(This article belongs to the Special Issue Digital Microfluidics for Liquid Handling and Biochemical Analysis)

Abstract

:
The two-phase flow in a microchannel consists of liquid–liquid and gas–liquid material components. The automatic recognition of flow patterns using deep learning approaches has been emerging. This study aimed to improve the recognition accuracy of flow patterns in the two-phase flow images. The different convolutional kernels in the GoogLeNet algorithm extracted the image features with different scales. In order to strengthen the important channel and spatial features, this paper proposes the combined five-layer Coord attention and GoogLeNet algorithm to enhance the accuracy of the new algorithm. The optimized algorithm model was derived from image datasets with different liquid–liquid two-phase flows (NaAlg–Oil, GaInSn–Water), and its accuracy was 95.09% in training and 98.12% in testing. This new model was also applied to predict the flow patterns, with a recognition accuracy of more than 97% in both the liquid–liquid and gas–liquid two-phase flows (water–soybean oil, water–lubricating oil, and argon–water).

1. Introduction

A flow system composed of mutually immiscible two-phase substances (with at least one phase being a fluid) is called a two-phase flow. In two-phase flow experiments, the continuous phase and the dispersed phase enter the same channel from different channels, presenting different flow patterns. At present, flow patterns play an important role in the fields of biomedicine, material synthesis, and aerospace [1,2,3]. In the study of two-phase flow images, there are many types of flow pattern; for instance, slug, dripping, and jetting are the flow patterns used to generate monodispersed droplets. The traditional recognition of flow patterns mainly relied on visual observation. Direct visual recognition is effective in low-speed videos, but it is not suitable for high-speed videos. The structure of the two-phase interface is complex, and the flow patterns may be converted in the video, causing some flow patterns to be misclassified. Considering the large number and the inconsistent quality of the flow pattern images, as well as the subjectivity of human observation, scholars have tried to use deep learning to identify flow patterns.
Convolutional neural networks (CNN)—one of the mainstream approaches in deep learning—can recognize and classify the flow patterns in research on two-phase flows. Many scholars have improved the feature extraction ability to obtain better network performance in CNN by increasing the network depth [4,5], enhancing the architecture of the convolution module [6,7,8], and adding new functional units [9]. The mobile network [10,11] and the shuffle network [12,13] increase the deep separable convolution to obtain algorithm models with lower computational cost and higher accuracy. The GoogLeNet algorithm [14] uses convolutional kernels of different sizes to extract the image features at different scales, enhancing the recognition and classification accuracy of the network model.
Adding an attention mechanism to a CNN can improve the recognition accuracy in the image identification. SENet [15] strengthens the important channel features and improves the recognition accuracy. CBAM [16], involving the spatial attention module in SENet, has better recognition and classification results. Coord [17] attention embeds the location information into the channel attention, enhancing the accuracy. LKA [18] embeds the self-attention mechanism into the large kernel convolution for extracting global information.
The research on the recognition algorithms of two-phase flow patterns has mainly focused on the architecture of network algorithms, the optimization of datasets, and the extension of algorithm models. VGG [19,20], ResNet [21,22], and GoogLeNet [23] are the most commonly used algorithms in flow pattern recognition due to their good performance. Some researchers have tried to establish algorithm models to recognize flow patterns using image datasets from different material components [24,25,26], synthetic algorithms [27], and data enhancement algorithms [28]. Nie, F. predicted the flow pattern of nitrogen–liquid nitrogen using a trained algorithm model that was extracted from the tetrafluoromethane–methane flow [23].
To further advance research in this field, this paper introduces an attention mechanism into the GoogLeNet algorithm, improving the recognition accuracy of flow patterns. The optimized model can predict the flow patterns of both liquid–liquid and gas–liquid two-phase flows.

2. Experiment and Image Dataset

Our two-phase flow experiments had two kinds of material components (NaAlg–oil, GaInSn–water), with two kinds of microchannels (convergent coaxial and vertical coaxial). The experimental parameters and images are listed in Table 1 and Table 2.
There were four flow patterns (slug, dripping, jetting, and others) in all experiments. The dispersed phases in the slug, dripping, and jetting flow patterns were in the shape of monodisperse droplets. The droplet length of the jetting pattern was obviously smaller than the microchannel width. The droplet length of the dripping pattern was 1.5 times smaller than the inner diameter of the microchannel. Meanwhile, the droplet length of the slug pattern was 1.5 times larger than the inner diameter of the microchannel. In addition to these three flow patterns, the other flow patterns in the two-phase flow were denoted as others.
A total of 24,860 images were collected from the raw videos of the experiments, captured by a high-speed camera. The collection rules for the image datasets were as follows: (1) 15 images with continuous frames were collected from each experiment video, (2) the size of each image was adjusted to be 224 × 224, and (3) similar features were included in each image, such as the channel architecture, the two-phase flows, etc.
The images in this paper had inconsistent sizes; however, the input image pixel in the deep learning algorithm was normalized to be 224 × 224. To match the requirements of the image pixel, the image size was transformed twice from the original pixels to 224 × 224 (Table 3).
In total, the slug pattern occupied 4.2% of the original image datasets. It was obvious that the number of images with a slug pattern was too small to affect the algorithm’s accuracy compared to the other flow patterns. Thus, the data enhancement algorithm was introduced to increase the percentage of slug patterns to 16.9%. As a result, the percentages of the other three flow patterns (dripping, jetting, and others) were 46.66%, 18.97%, and 17.47%, respectively.

3. GoogLeNet Algorithm

In general, three-dimensional matrices describe the total image information in deep learning algorithms. The most popular algorithms adopt a matrix of 224 × 224 × 3 as the input parameter, in which 224 × 224 is the size of the image and 3 is the number of red, green, and blue channels. In the processer, many convolution and pooling layers reduce the size of channels (feature maps) and increase the number of channels simultaneously. As a result, the one-dimensional matrix is transformed from the three-dimensional matrices and is set to the softmax classifier, calculating the multi-classification output.
This paper studied the efficacy of four deep learning algorithms in the pattern recognition of two-phase flows.
GoogLeNet [14]: The GoogLeNet algorithm (Figure 1) is 22 layers in depth for parameters or 27 layers in depth for pooling counting. Its first four layers are the convolutional layers and the pooling layers. Beneath the algorithm layer, the GoogLeNet algorithm creates the inception module (Figure 1). For example, there are two layers of the inception module in the feature maps of 28 × 28 and 7 × 7, and five layers of the inception module in the feature map of 14 × 14.
The inception module connects the multiple convolution kernels and the maximum pooling in parallel. It consists of three convolutions with different sizes and a maximum pooling layer (Figure 2). The calculation of the multiple convolution kernels increases explosively as the number of algorithm layers increases. The inception module introduces the 1 × 1 convolution to reduce the computational cost. The parallel 1 × 1, 3 × 3, and 5 × 5 convolutions can extract richer features in a layer, resulting in higher recognition accuracy.
VGG 16 [5]: The Visual Geometry Group Network (VGG16) has a total of 13 convolutional layers and 3 fully connected layers. All of the convolution and pooling layers adopt 3 × 3 convolutional kernels and 2 × 2 pooling kernels.
ViT [29]: Vision Transformer (ViT) introduces the Transformer module. The input of the Transformer module is a one-dimensional matrix transformed from three-dimensional matrices in the algorithm. The whole image being flattening into a one-dimensional matrix addresses the issue of huge computational cost. ViT divides the image into 16 windows and flattens it into a one-dimensional matrix to reduce the amount of calculation.
Swin-T [30]: Swin-Transformer (Swin-T) also applies the Transformer module. Unlike ViT, which calculates the self-attention in the image, Swin-T calculates the Transformer in each small window first, and then merges the small windows into large windows. After the three iterations of window merging and calculation, the output of the Transformer is sent into the fully connected layer and the softmax layer.
For comparison, VGG16, GoogLeNet, ViT, and Swin-T were trained to recognize the flow patterns after 50 iterations based on our image datasets in the liquid–liquid two-phase flows (Figure 3). The green solid line represents VGG16, with a training accuracy rate of 58.16%. The yellow solid line represents ViT, with a training accuracy rate of 86.71%. The blue solid line represents Swin-T, with a training accuracy rate of 91.74%. The red solid line represents GoogLeNet, with a training accuracy rate of 94.83%. In all experimental results, GoogLeNet showed the best recognition accuracy for the two-phase flow patterns.

4. Coordinate Attention

In order to improve the recognition accuracy of GoogLeNet beyond 94.83%, this paper implanted an attention mechanism into the GoogLeNet algorithm.
The principle of an attention mechanism is to locate the interesting information and suppress the useless information by changing the weights of different areas. Attention has been widely used in computer vision and natural language processing.
The following four common attention were discussed:
Coord: Coordinate attention (Coord) embeds the positional information into the channel attention, and strengthens the channel and spatial features (Figure 4).
As shown above, Coord attention implements two pooling kernels, (H, 1) or (1, W), to encode spatial information in the horizontal channel and the vertical channel, respectively. Equations (1) and (2) give the calculation equations:
Z c h ( h ) = 1 W 0 i < W x c ( h , i )
Z c W ( w ) = 1 H 0 i < H x c ( j , w )
where xc is the input of the c-th dimension, Z(h) is the H-dimensional output of the c-th channel, and Z(w) is the W-dimensional output of the c-th channel.
The contact layer and 1 × 1 convolution compress the channel in the spatial dimension and encode the spatial information in the vertical and horizontal directions through the batch-norm. The output is transformed into a pair of feature maps by the convolution transformation function. The convolution transformation function is
F = δ ( F 1 ( Z h , Z w ) )
where F is RC/r*(H + W), and δ is the nonlinear activation function.
Finally, the spatial features in the horizontal and vertical channels are calculated separately through 1 × 1 convolution, and then the features of the convolutions are put together to recalculate the weights. The formulae of output Y are
G h = σ ( F h ( f h ) )
G w = σ ( F w ( f w ) )
y c ( i , j ) = x c ( i , j ) × G c h ( i ) × G c w ( j )
where fh is RC/r*H, and fw is RC/r*W.
Coord attention accounts for both the channel-to-channel relationships and the location information. It captures not only the information across channels, but also the direction-aware and position-sensitive information, which can accurately locate and identify the target areas.
CBAM: The convolutional block attention module (CBAM) is divided into channel and spatial modules. The channel attention module compresses the feature map in the spatial dimension. The spatial attention module compresses the channel in the spatial dimension. CBAM sequentially generates an attention map in two independent dimensions (channel and space).
LKA: Large kernel attention (LKA) introduces self-attention into the large convolution kernels. The convolution of large kernel size is decomposed into the depthwise convolution, the depthwise empty convolution, and the channel convolution. LKA establishes the correlation of each point, generating the attention map in the large kernel convolution, which realizes the adaptability of the channel dimension and the spatial dimension.
SENet: Squeeze-and-Excitation Networks (SENet) are divided into squeeze and excitation modules. The squeeze module performs the feature compression in the spatial dimension, turning the 2D feature channel into a real number. The excitation module calculates the weights and correlations between each feature channel.

5. Optimized Algorithm Results and Discussion

5.1. Optimized Algorithm Architecture

Four kinds of attention (Coord, CBAM, LKA, and SENet) were introduced into the GoogLeNet algorithm for the training and validation datasets. The innovative architectures possessed five layers of attention to be embedded in each map, with sizes of 224 × 224, 56 × 56, 28 × 28, 14 × 14, and 7 × 7. Figure 5 illustrates one architecture of the four attention, and Table 4 gives the results of their recognition accuracy. Compared with the three other attention, the accuracy of Coord attention was the best. This new GoogLeNet+5 Coord attention algorithm was proposed to recognize flow patterns.
GoogLeNet focuses on the local features extracted by the convolution kernels with different sizes; however, GoogLeNet is not good at global feature extraction. Coord attention can extract global features and also strengthen the important spatial and channel features. Coord attention compensates for the lack of global, spatial and channel features in GoogLeNet due to its computational reduction. Introducing Coord attention into the GoogLeNet algorithm improved the performance in recognizing the flow patterns, with a high accuracy and low loss in the training and validation steps.
In our two-phase flow images, the GoogLeNet algorithm easily extracted the local features, especially for the droplet features. Additionally, Coord attention was supplemented to extract the global features of images, such as the background and microchannel features. The innovative GoogLeNet+5 layers of the attention algorithm converged the local features of two-phase flows and the global features of the background and microchannels, which distinguished the background features clearly to recognize the flow patterns with a higher accuracy.

5.2. Training and Testing Results

The dataset was separated into model and test datasets at a ratio of 8:2, and the model dataset was also separated into training and validation datasets at a ratio of 8:2. For the training dataset, setting the batch size to 32, the learning rate to 0.0001, and the iteration number to 50, the recognition results of GoogLeNet and GoogLeNet+5 Coord are plotted together in Figure 6.
The blue solid curve represents the training results of GoogLeNet, with an accuracy of 94.83% and loss of 0.1245. The red solid curve represents the training results of GoogLeNet+5 Coord, with an accuracy of 95.09% and loss of 0.1222. This indicates that the new algorithm, GoogLeNet+5 Coord, had a higher training accuracy and lower training loss than the traditional GoogLeNet algorithm—by 0.26% and −0.0023, respectively.
Figure 7 presents the validation results of GoogLeNet+5 Coord and GoogLeNet. The blue solid curve represents the validation result of GoogLeNet, with an accuracy of 98.67% and loss of 0.03574. The green solid curve represents the validation result of GoogLeNet+5 Coord, with an accuracy of 98.87% and loss of 0.03221. The new GoogLeNet+5 Coord had better validation results for accuracy (+0.2%) and loss (−0.00353) than the traditional GoogLeNet algorithm.
After the establishment of the algorithm model, the remaining 20% of the data were used in the testing dataset to test the model precision of GoogLeNet+5 Coord and GoogLeNet (Figure 8). For 50 iterations with the testing dataset, the average recognition accuracy of all images when applying GoogLeNet+5 Coord was 98.12%, which was about 0.29% higher than that of GoogLeNet.

5.3. Prediction Results

The optimized algorithm model with a higher accuracy (>98%) was derived from the image datasets with liquid–liquid two-phase flows (NaAlg–oil, GaInSn–water).
This paper extended the model of GoogLeNet+5 Coord to predict the flow patterns in the different gas–liquid and liquid–liquid two-phase components. Similarly, the datasets contained 600 images with the four flow patterns (Table 5). The average accuracy was 97.65% in the prediction of gas–liquid flow patterns.
Figure 9 depicts the prediction accuracy of the two algorithms in both the liquid–liquid and gas–liquid two-phase flows. GoogLeNet+5 Coord had better performance than GoogLeNet. The new algorithm model could accurately identify the four flow patterns, with an average accuracy of 97.83%. GoogLeNet had poor accuracy in identifying the slug flow pattern, with an average accuracy of 84.25%. This shows that GoogLeNet+5 Coord had higher accuracy and better consistency than GoogLeNet.
In our two-phase flow images, the GoogLeNet algorithm easily extracted the local features, especially for the small droplet features. For example, for jetting, the largest convolution of GoogLeNet was 7 × 7, which was approximate to the size of a droplet. Therefore, the prediction accuracy for jetting in GoogLeNet was higher than in GoogLeNet+5 Coord. Moreover, the convolution of 7 × 7 was much smaller than the droplet size of the slug flow pattern, and the recognition accuracy of the slug pattern in prediction was poor. Coord attention was supplemented to extract the global features of images, such as the background features, large droplet features, microchannel features, etc. The innovative GoogLeNet + 5 layers of attention algorithm converged the local features of the two-phase flows and the global features of the large droplets, the background, and the microchannels to accurately predict the four flow patterns (Figure 10).

6. Conclusions

This paper researched flow pattern recognition from image datasets of gas–liquid and liquid–liquid two-phase flows.
1. Compared with the VGG16, ViT, and Swin-T algorithms, the GoogLeNet algorithm had a higher accuracy in recognizing and classifying flow patterns.
2. Different attention were introduced to improve the recognition accuracy, and it was found that the optimal algorithm was GoogLeNet+5 Coord, which strengthened the important channel and spatial features and extracted the two-phase flow features simultaneously.
3. The optimized GoogLeNet+5 Coord algorithm was trained from the data of different liquid–liquid two-phase flows, and it could predict the liquid–liquid and gas–liquid two-phase flow patterns with a high accuracy of more than 97%.
4. The optimized algorithm model was a normalized model for flow pattern recognition in both liquid–liquid and gas–liquid two-phase flows.

Author Contributions

Conceptualization, J.Z. and Z.W.; methodology, X.W.; software, X.W.; validation, J.Z. and Z.W.; formal analysis, X.W.; investigation, J.Z.; resources, Z.W.; data curation, X.W.; writing—original draft preparation, X.W.; writing—review and editing, J.Z.; visualization, X.W.; supervision, J.Z.; project administration, J.Z.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Haase, S.; Marschner, S.; Ayubi, M.-M.; Lange, M. Gas-liquid flow in small channels: Artificial neural network classifiers for flow regime prediction. Chem. Eng. Process.-Process. Intensif. 2022, 180, 108687. [Google Scholar] [CrossRef]
  2. Wu, Z.; Sundén, B. Liquid-liquid two-phase flow patterns in ultra-shallow straight and serpentine microchannels. Heat Mass Transf. 2019, 55, 1095–1108. [Google Scholar] [CrossRef]
  3. Qian, J.-Y.; Li, X.-J.; Wu, Z.; Jin, Z.-J.; Sunden, B. A comprehensive review on liquid–liquid two-phase flow in microchannel: Flow pattern and mass transfer. Microfluid. Nanofluid. 2019, 23, 116. [Google Scholar] [CrossRef]
  4. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  5. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 16, pp. 770–812. [Google Scholar]
  7. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar] [CrossRef]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  9. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  10. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  11. Sandler, M.; Howard, A.G.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; Volume 10, pp. 4510–4520. [Google Scholar]
  12. Howard, A.G.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; Volume 10, pp. 1314–1324. [Google Scholar]
  13. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; Volume 24, pp. 6848–6856. [Google Scholar]
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A.; Liu, W.; et al. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; Volume 26, pp. 1–9. [Google Scholar] [CrossRef]
  15. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; Volume 10, pp. 7132–7141. [Google Scholar]
  16. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Volume 2, pp. 3–19. [Google Scholar]
  17. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; Volume 86, pp. 13713–13722. [Google Scholar]
  18. Guo, M.H.; Lu, C.Z.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Visual attention network. arXiv 2022, arXiv:2202.09741. [Google Scholar]
  19. Roman, A.J.; Kreitzer, P.J.; Ervin, J.S.; Hanchak, M.S.; Byrd, L.W. Flow pattern identification of horizontal two-phase refrigerant flow using neural networks. Int. Commun. Heat Mass Transf. 2016, 71, 254–264. [Google Scholar] [CrossRef]
  20. Tong, W.G.; Pang, X.C.; Zhu, G.H. Flow pattern identification method of gas-liquid two-phase flow based on convolutional neural networks. Chin. J. Syst. Simul. 2021, 33, 883–884. [Google Scholar]
  21. Zheng, Q.; Xu, Y.; Zhang, P.; Bian, J.; Wang, F. Identification of gas-liquid two-phase flow regime in pipelines with low liquid holdup based on ResNet1D-34. Flow Meas. Instrum. 2022, 88, 102249. [Google Scholar] [CrossRef]
  22. Liang, F.; Hang, Y.; Yu, H.; Gao, J. Identification of gas-liquid two-phase flow patterns in a horizontal pipe based on ultrasonic echoes and RBF neural network. Flow Meas. Instrum. 2021, 79, 101960. [Google Scholar] [CrossRef]
  23. Nie, F.; Wang, H.; Song, Q.; Zhao, Y.; Shen, J.; Gong, M. Image identification for two-phase flow patterns based on CNN algorithms. Int. J. Multiph. Flow 2022, 152, 104067. [Google Scholar] [CrossRef]
  24. Poletaev, I.; Tokarev, M.P.; Pervunin, K.S. Bubble patterns recognition using neural networks: Application to the analysis of a two-phase bubbly jet. Int. J. Multiph. Flow. 2020, 126, 103194. [Google Scholar] [CrossRef]
  25. Liu, L.; Bai, B. Flow regime identification of swirling gas-liquid flow with image processing technique and neural networks. Chem. Eng. Sci. 2019, 199, 588–601. [Google Scholar] [CrossRef]
  26. Du, M.; Yin, H.; Chen, X.; Wang, X. Oil-in-water two-phase flow pattern identification from experimental snapshots using convolutional neural network. IEEE Access 2019, 73, 6219–6225. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Azman, A.N.; Xu, K.W.; Kang, C.; Kim, H.B. Two-phase flow regime identification based on the liquid-phase velocity information and machine learning. Exp. Fluids 2020, 61, 212. [Google Scholar] [CrossRef]
  28. Shen, C.; Zheng, Q.; Shang, M.; Zha, L.; Su, Y. Using deep learning to recognize liquid–liquid flow patterns in microchannels. AIChE J. 2020, 66, e16260. [Google Scholar] [CrossRef]
  29. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  30. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; Volume 10, pp. 9992–10002. [Google Scholar]
Figure 1. GoogLeNet architecture.
Figure 1. GoogLeNet architecture.
Micromachines 14 00462 g001
Figure 2. Inception architecture.
Figure 2. Inception architecture.
Micromachines 14 00462 g002
Figure 3. Training accuracy of four algorithms (VGG, ViT, Swin-T, and GoogLeNet).
Figure 3. Training accuracy of four algorithms (VGG, ViT, Swin-T, and GoogLeNet).
Micromachines 14 00462 g003
Figure 4. Coordinate attention architecture.
Figure 4. Coordinate attention architecture.
Micromachines 14 00462 g004
Figure 5. Architecture of GoogLeNet+5 Coord attention.
Figure 5. Architecture of GoogLeNet+5 Coord attention.
Micromachines 14 00462 g005
Figure 6. Accuracy and loss of the training dataset when applying two algorithms: (a) training accuracy; (b) training loss.
Figure 6. Accuracy and loss of the training dataset when applying two algorithms: (a) training accuracy; (b) training loss.
Micromachines 14 00462 g006
Figure 7. Accuracy and loss of the validation dataset when applying two algorithms: (a) validation accuracy; (b) validation loss.
Figure 7. Accuracy and loss of the validation dataset when applying two algorithms: (a) validation accuracy; (b) validation loss.
Micromachines 14 00462 g007
Figure 8. Testing accuracy of GoogLeNet+5 Coord and GoogLeNet.
Figure 8. Testing accuracy of GoogLeNet+5 Coord and GoogLeNet.
Micromachines 14 00462 g008
Figure 9. Prediction accuracy of two-phase flows.
Figure 9. Prediction accuracy of two-phase flows.
Micromachines 14 00462 g009
Figure 10. Local features and global features of slug and jetting images.
Figure 10. Local features and global features of slug and jetting images.
Micromachines 14 00462 g010
Table 1. Experiment parameters of two-phase flows.
Table 1. Experiment parameters of two-phase flows.
Material ComponentsMicrochannel TypeConvergence AngleDispersed PhaseContinuous PhaseExperimental
Groups
NaAlg
–Oil
Convergent coaxial NaAlg
(10~480 mL/h)
Soybean oil (1~10 mL/h)773
Oil–
NaAlg
Convergent coaxial Soybean oil (1~250 mL/h)NaAlg
(1~10 mL/h)
734
NaAlg
–Oil
Vertical coaxial--NaAlg
(10~480 mL/h)
Soybean oil (1~10 mL/h)658
Oil–
NaAlg
Vertical coaxial --Soybean oil (1~250 mL/h)NaAlg
(1~10 mL/h)
439
GaInSn–waterVertical coaxial--GaInSn (7~108 mL/h)Water
(36~900 mL/h)
132
Table 2. The definition of flow patterns with different images.
Table 2. The definition of flow patterns with different images.
Flow PatternFeaturesExperiment ImagesQuantity of Images
SlugMonodisperse droplets with a bullet-like or plunger-like shape, and the droplet length is 1.5 times larger than the inner diameter of the microchannelMicromachines 14 00462 i001Micromachines 14 00462 i002Micromachines 14 00462 i00313624
DrippingMonodisperse droplets with an ellipsoid or spherical shape, and the droplet length is 1.5 times smaller than the inner diameter of the microchannelMicromachines 14 00462 i004Micromachines 14 00462 i005Micromachines 14 00462 i006110,385
JettingMonodisperse droplets with a nearly spherical shape, and the droplet length is significantly smaller than the inner diameter of the microchannel, with fast generation frequency and a stretching neckMicromachines 14 00462 i007Micromachines 14 00462 i0081Micromachines 14 00462 i00914366
OthersOther flow patternsMicromachines 14 00462 i010Micromachines 14 00462 i0111Micromachines 14 00462 i01214717
1 Represented blurred image.
Table 3. Original pixels transformed into uniform pixels.
Table 3. Original pixels transformed into uniform pixels.
SlugDrippingJettingOthers
Images of original pixelsMicromachines 14 00462 i013Micromachines 14 00462 i014Micromachines 14 00462 i015Micromachines 14 00462 i016
Images of uniform pixelsMicromachines 14 00462 i017Micromachines 14 00462 i018Micromachines 14 00462 i019Micromachines 14 00462 i020
Table 4. The accuracy of the GoogLeNet+5 layers with different attention.
Table 4. The accuracy of the GoogLeNet+5 layers with different attention.
AlgorithmTraining Accuracy (%)Validation Accuracy (%)
GoogLeNet94.8398.67
GoogLeNet+5 Coord95.0998.87
GoogLeNet+5 CBAM94.9798.72
GoogLeNet+5 LKA94.7198.54
GoogLeNet+5 SENet95.0498.67
Table 5. Experiments for flow pattern prediction.
Table 5. Experiments for flow pattern prediction.
Material ComponentsDispersed PhaseContinuous PhaseExperimental
Groups
Images
Oil–waterVegetable oilWater59Micromachines 14 00462 i021Micromachines 14 00462 i022Micromachines 14 00462 i023Micromachines 14 00462 i024
Oil–waterLubricating oilDeionized water52Micromachines 14 00462 i025Micromachines 14 00462 i026Micromachines 14 00462 i027
Argon–waterArgonWater93Micromachines 14 00462 i028Micromachines 14 00462 i029Micromachines 14 00462 i030Micromachines 14 00462 i031
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Wei, X.; Wang, Z. The Recognition Algorithm of Two-Phase Flow Patterns Based on GoogLeNet+5 Coord Attention. Micromachines 2023, 14, 462. https://doi.org/10.3390/mi14020462

AMA Style

Zhang J, Wei X, Wang Z. The Recognition Algorithm of Two-Phase Flow Patterns Based on GoogLeNet+5 Coord Attention. Micromachines. 2023; 14(2):462. https://doi.org/10.3390/mi14020462

Chicago/Turabian Style

Zhang, Jinsong, Xinpeng Wei, and Zhiliang Wang. 2023. "The Recognition Algorithm of Two-Phase Flow Patterns Based on GoogLeNet+5 Coord Attention" Micromachines 14, no. 2: 462. https://doi.org/10.3390/mi14020462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop