Next Article in Journal
Capacitive Effect and Electromagnetic Coupling on Manganin Gauge Limiting the Bandwidth for Pressure Measurements under Shock Conditions
Previous Article in Journal
Patient-Specific Preictal Pattern-Aware Epileptic Seizure Prediction with Federated Learning
Previous Article in Special Issue
Pedestrian Trajectory Prediction for Real-Time Autonomous Systems via Context-Augmented Transformer Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

End-to-End Bubble Size Distribution Detection Technique in Dense Bubbly Flows Based on You Only Look Once Architecture

1
College of Control Science and Engineering, Zhejiang University, Hangzhou 310027, China
2
Luzhou Laojiao Co., Ltd., Luzhou 646000, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6582; https://doi.org/10.3390/s23146582
Submission received: 9 June 2023 / Revised: 16 July 2023 / Accepted: 19 July 2023 / Published: 21 July 2023
(This article belongs to the Collection Computer Vision Based Smart Sensing)

Abstract

:
Accurate measurements of the bubble size distribution (BSD) are crucial for investigating gas–liquid mass transfer mechanisms and describing the characteristics of chemical production. However, measuring the BSD in high-density bubbly flows remains challenging due to limited image algorithms and high data densities. Therefore, an end-to-end BSD detection method in dense bubbly flows based on deep learning is proposed in this paper. The bubble detector locates the positions of dense bubbles utilizing objection detection networks and simultaneously performs ellipse parameter fitting to measure the size of the bubbles. Different You Only Look Once (YOLO) architectures are compared, and YOLOv7 is selected as the backbone network. The complete intersection over union calculation method is modified by the circumferential horizontal rectangle of bubbles, and the loss function is optimized by adding L 2 constraints of ellipse size parameters. The experimental results show that the proposed technique surpasses existing methods in terms of precision, recall, and mean square error, achieving values of 0.9871, 0.8725, and 3.8299, respectively. The proposed technique demonstrates high efficiency and accuracy when measuring BSDs in high-density bubbly flows and has the potential for practical applications.

1. Introduction

Bubbly flow is a common gas–liquid two-phase flow pattern which widely exists in water conservancy, petroleum, the chemical industry, the nuclear industry, and other fields. For example, the shape and size distributions of bubbles have a high correlation with the performance of mineral froth flotation [1]. The morphology of bubbles in Chinese spirits is closely related to alcohol concentration and quality [2]. Monitoring bubble parameters in hysteroscopic images is especially necessary to diagnose gas embolisms [3]. Detailed knowledge of bubble characteristics and dynamics is essential for process optimization. The interaction between the bubbles and the liquid determines the quantity of transferred mixing energy. Having an accurate measurement of the bubble size distribution (BSD) is very important for numerical flow modeling, the study of the gas–liquid mass transfer mechanism, and the design and control of gas–liquid multiphase reactors.
There are different BSD measurement methods for fluid mediums in the existing literature. They can be divided into intrusive and non-intrusive methods. Intrusive methods, such as capillary suction probes [4], conductivity probes [5], optical fiber probes [6], and wire-mesh sensors [7], are usually limited to a single or a few bubbles. Thus, the results may be biased by the applied measurement technique. Intrusive methods can be applied to a broader range of setups and with different liquids. Non-intrusive methods include phase doppler anemometry [8], interferometric particle imaging [9], and digital image analysis (DIA). Among non-invasive methods, direct imaging is the most commonly used since multiple objects can be observed simultaneously and the equipment is relatively cheap and flexible.
Traditional image processing techniques usually start with boundary extraction—either applying global or local thresholds—background subtraction, or edge detection by analyzing local intensity gradients. In the second stage, the boundary of the object, which can be individual bubbles or a cluster of bubbles, is segmented into curved segments of the bubbles. This can be achieved through seed-point extraction or breakpoint-based methods. Seed-points are defined as the centers of overlapping objects and are used to determine the number of bubbles in a cluster [10]. Different techniques are used, namely bounded erosion with elliptical distance transform [11], ultimate erosion [12], and fast radial symmetry transform [13]. The watershedding technique [14] and polygonal approximation method [15] are proposed to separate bubble objects into solitary bubbles and overlapping bubbles. Breakpoint methods segment the contours of clusters into contours of individual objects by detecting points where contours intersect. This can be done by analyzing the convexity of the contour [16] or by a combination of boundary curvature and intensity gradient analysis [17]. Several methods for reconstructing overlapping bubbles are proposed. An improved Hough transform algorithm can be used to reconstruct the contours of bubbles [18]. Ellipse parameters are estimated as bubble parameters using a stable direct least square fitting method [15].
A major problem with traditional imaging techniques is that the parameters of the different stages and even the workflow itself strongly depend on the experimental setup in which the images were taken. Experimental variables such as the distance between bubbles and the camera or the background light intensity significantly affect the applicability of these techniques. However, these variables are often limited by experimental conditions. Therefore, the potential for generalizing workflows is limited, requiring extensive manual adaptations and expertise in image processing.
Deep learning has become a prevailing machine vision technique due to the development of deep convolutional neural networks (CNN) in recent years. By comprehensively comparing the performance of various bubble detection methods when processing industrial image data, it is proved that the CNN-based method is superior to the circular Hough transform, concentric circular arrangements, and boosting-based detection methods [19]. Based on the deep learning algorithm, bubble center detection and synthetic bubble image generation are proposed to determine overlapping, blurred, and non-spherical bubble images [20]. A convolutional denoising auto-encoder architecture was introduced to recognize bubble patterns in images and identify their geometric parameters. The training dataset used synthetic images similar to real photos collected in experiments [21]. BubCNN employed a faster region-based CNN (RCNN) detector to localize bubbles and a shape regression CNN to predict bubble shape parameters [22]. Compared with state-of-the-art image processing procedures, BubCNN shows better generalization abilities. While the accuracy of BubCNN is high when the gas holdup is below 2%, its accuracy drops significantly as the gas holdup increases. Cerqueira and Paladino [23] developed a CNN-based method to identify bubbles and reconstruct bubble shapes in millimeter bubbly flows and obtained accurate results at gas holdups below 9%. A framework combining a deep edge-aware network and a marker-controlled watershed algorithm was proposed for extracting bubble parameters from hysteroscopy images. The proposed edge-aware network consists of an encoder–decoder architecture for bubble segmentation and a contour branch that is supervised by edge losses [3].
Researchers have also made great efforts in BSD detection in froth flotation. McGill bubble size analyzers (BSA) [24] are known to be the most acceptable techniques in both batch and industrial conditions. The Anglo Platinum Bubble Sizer was developed by the University of Cape Town in collaboration with Anglo Platinum as a modification of McGill BSA. Ma et al. [25] used a linear relationship between the bubble size and the area of the bright centroid to improve BSD characterization in the spherical regime in the presence of dense binding clusters. Bubble sampling is also critical to reduce the presence of clusters. Azgomi et al. [26] measured bubble size in a laboratory-scale flotation column which was designed with an expansion in the section. The results showed that the gas holdup decrease in the sampling point was effective in reducing the bubbles observed in the visual field. A new deflecting system that allowed a fraction of the sampled bubbles to be photographed was designed to reduce clusters of bubbles in the visual field [27]. An online BSD monitoring scheme was proposed by incorporating a multiscale-deblurring full convolutional network (FulConNet) (MsD) and a multistage jumping feature-fused FulConNet (MsJ), having the potential of online identification of the health state of the flotation process operations [28].
Although the current deep learning technology has improved the accuracy and robustness of bubble detection methods, it is worthwhile to mention that some problems remain. Firstly, the characteristics of bubbly flows with high gas holdup, such as low transparency, overlapping bubbles, and significant variations in bubble scale, impede the practical implementation of conventional bubble detection methods. Deep-learning-based methods in the literature are still limited to applications in gas–liquid flow with low gas holdups. Secondly, the current CNN models are typically designed as two-stage methods involving segmentation and reconstruction, which require a huge amount of computation and can impact the efficiency of the detection process.
This work proposes an end-to-end BSD detection method for dense bubbly flows. The proposed model optimizes the output structure and loss function, allowing it to effectively locate bubbles and determine size parameters through ellipse fitting simultaneously. The design of the loss function takes into account both localization loss and size fitting loss, thereby improving the accuracy of both bubble localization and size regression. Efficient objection detection model YOLO series models are researched and adopted as backbone models. This paper is structured as follows. In Section 2, we give a detailed description of the proposed method. In Section 3, we present the experimental results and analysis. Section 4 discusses the advantages and disadvantages of the proposed method. Finally, the conclusions are presented in Section 5.

2. Methodology

2.1. Overview

In contrast to previous tea-stage methods, an end-to-end BSD detection scheme based on deep learning is proposed in this paper, as shown in Figure 1. This scheme uses the objection detection framework to realize bubble location and size detection in the dense flow simultaneously. The whole process is divided into two stages of training and detection. During training, a bubble detector based on a deep CNN model is trained on public bubble datasets [29]. The model parameters are updated via backpropagation. During detection, the well-trained detector model takes bubble images as input, then performs the forward propagation, and finally outputs the bubble parameters.
In the field of computer vision, object detection includes two tasks, namely object location and object classification. Object classification is a classification problem to determine what category the target belongs to. Object location is essentially a regression problem that determines the coordinate position and size (usually the length and width of the bounding box) of the targets in the image. In this paper, the proposed method is only responsible for detecting a single category which is the “bubble” and does not distinguish the properties of bubbles (such as bubble shape, bubble color, etc.). Hence, the detector model ignores the classification requirement. BSD detection is essentially a regression prediction of the location and size parameters of the bubbles.
In many processes, the shape of rising gas bubbles can be approximated by an ellipsoid [22]. In a 2D projection, ellipsoids become ellipses, which are described mathematically by
x x c cos ( θ ) + y y c sin ( θ ) 2 a 2 + x x c sin ( θ ) + y y c cos ( θ ) 2 b 2 = 1
where x c , y c is the ellipse center coordinate, and a , b are the semi-major and semi-minor axes of the ellipse, respectively. θ is the anticlockwise rotation angle. It is worth noting that the location parameters refer to x c , y c , and the size parameters refer to a , b , θ . The detailed processing tricks are explained below.

2.2. BSD Detector Architecture

The proposed BSD detection technique is based on the You Only Look Once (YOLO) algorithm [30,31,32]. YOLO series models are state-of-the-art models for real-time object detection with a fast network architecture, effective feature integration methods, and robust loss functions. YOLO divides the image into grids and predicts bounding boxes and probabilities for each grid simultaneously.
The one-stage BSD detector architecture is shown in Figure 2, including a backbone, a neck, and multi-head modules. The backbone is the feature extraction network with stacked convolution, pooling, and other operation layers, such as ResNet-50 [33], Darknet53 [30], CSPDarknet53 [31], etc. CNN modules in the backbone finally downsample the input by 32.
Nick is composed of several bottom-up paths and several top-down paths to collect feature maps from different stages, such as SPP [34], FPN [35], etc. To enhance the detection probability of different bubble targets, upsampling is carried out twice. This operation contributes to obtaining more meaningful semantic features and finer-grained information from the earlier feature map.
Multi-head detection is utilized on three feature maps of varying scales, downsampled by factors of 8, 16, and 32, respectively. The backbone feature extractor is augmented with multiple convolutional layers. The last layer outputs a three dimensional tensor encoding ellipse parameters and confidence. Figure 3 illustrates that each detection output tensor is encoded as S × S × ( B × ( 5 + 1 ) ) . This encoding is used because each output contains five ellipse parameters and one confidence parameter. S is the grid size of the feature map for each sub-detection path, and B is the anchor number for each path. It is worth noting that t x , t y , t a , t b are the offsets on the feature map instead of the final results, which is explained below.

2.3. Ellipse Location and Fitting

The initial approach of object detection typically employs the mean square error (MSE) for direct regression on the coordinates of the center point, as well as the height and width of the bounding box. Anchor-based methods estimate the corresponding offsets. Referring to the bounding box prediction method in YOLOv5 [36], the proposed model also utilizes the anchor-based strategy to detect bubbles. The difference lies in the design, where the detection network indirectly predicts four coordinates offsets ( t x , t y , t a , t b ) for each bubble, as shown in Figure 4. The final result is obtained using the following equations.
x c = 2 σ t x 0.5 + c x
y c = 2 σ t y 0.5 + c y
a = p a 2 σ t a 2
b = p b 2 σ t b 2
b w = 2 a 2 cos 2 ( θ ) + b 2 sin 2 ( θ )
b h = 2 b 2 cos 2 ( θ ) + a 2 sin 2 ( θ )
where c x , c y is the cell offset from the top left corner of the feature map. p a , p b are the prior semi-major and semi-minor axes of bubbles which are calculated by K-means [37]. In order to calculate intersection over union (IoU), the circumscribed horizontal rectangle of the rotated ellipse is used to replace the bounding box. b w and b h refer to the width and height of the circumscribed horizontal rectangle. All of the four offset parameters are predicted by a sigmoid function σ ( ) . The paper employs the pixel as the unit of x c , y c , a , b and radians as the unit of θ .

2.4. Loss Function Design

As described above, the backpropagation algorithm employs a loss function to constrain and guide the optimization of the model. We design a comprehensive loss function that includes location loss, regression loss, and confidence loss. Location loss L l o c is used to guide optimization of parameters t x and t y , regression loss L r e g is used to fit ellipse parameters t a , t b , and θ , and confidence loss L c o n f is used to evaluate an objectness score for each bubble. The entire loss function L o s s is defined as
L o s s = λ l L l o c + λ r L r e g + λ c L c o n f
L l o c = 1 C I o U
L r e g = i = 0 S 2 j = 0 B I i j o b j ( a i j a ^ i j ) 2 + ( b i j b ^ i j ) 2 + ( θ i j θ ^ i j ) 2
L c o n f = 1 N i = 0 S 2 j = 0 B o i j ln c o n f i j + ( 1 o i j ) ln 1 c o n f i j
where λ l , λ r , λ c are constant factors. We use complete-IoU (CIoU) [38] to calculate location loss. Here, the bounding box is replaced by x c , y c , b w , b h . Regression loss is calculated by the mean square error loss. I i o b j denotes if an object appears in cell i, and I i j o b j denotes that the jth target in cell i is responsible for that detection. Confidence loss is calculated by binary cross-entropy loss. a i j , b i j , θ i j are the predicted values, and a ^ i j , b ^ i j , θ ^ i j are the true values. o i j 0 , 1 refers to the CIoU of the predicted target bounding box and real target bounding box, c o n f i j is the prediction confidence obtained by sigmoid function, and N is the total number of positive and negative samples.

3. Experiments

3.1. Dataset

All images in the dataset were synthetically generated using the public BubGAN [29] tool. BubGAN can be used for synthetic bubbly flow generation with customized bubbly flow boundary conditions. A total of one million synthetic bubbles are stored in the MillionBubble database with known bubble properties in terms of aspect ratio, rotation angle, circularity, and edge ratio.
In this work, 100,000 synthetic bubbly flow images were generated by assembling single bubbles in MillionBubble on an image background canvas with a size of 600 × 600 pixels. A total of 8000 images in the dataset are used as the training set, and the remaining 2000 images are used as the test set. The key parameters for generating images are bubble center coordinate x c , y c , semi-major and semi-minor axes a , b , rotation angle θ , and void fraction. The distribution of the parameters is shown in Figure 5.
The center coordinates are subject to a random uniform distribution within the size of the canvas. The rotation angles are normally distributed. The semi-major axis and the semi-minor axis are mainly randomly distributed in 5 , 40 pixels. The images have a resolution of around 25 pixels per mm. Bubbles are generally easy to detect in spherical regimes, while the detectability becomes more challenging when large bubbles coexist with small bubbles [39]. The rotation angle follows a Gaussian distribution, and its range is π / 2 , π / 2 . The void fraction is defined by the volume of all bubbles divided by the total volume, which is randomly and uniformly distributed in 0.01 , 0.1 . The larger the void fraction is, the denser the bubbles become. Note that the void fraction in the test set is chosen from 0.01 , 0.02 , 0.03 , 0.04 , 0.05 , 0.06 , 0.07 , 0.08 , 0.09 , 0.1 , and 200 images are generated under each value.

3.2. Evaluation Metrics

In this work, the flowing several metrics are used to evaluate the trained detectors.
(1)
Precision: The fraction of relevant instances among the retrieved instances
Precision = True Positive True Positive + False Positive
(2)
Recall: The fraction of relevant instances that were retrieved
Recall = True Positive True Positive + False Negtaive
(3)
AP50: the average precision (AP) value calculated at the threshold of 50% for detection evaluation. Specially, average precision is equivalent to mean average precision (mAP) due to a single-class detection task.
(4)
F1 score: A harmonic average of precision and recall
F 1 = 2 × Precision × Recall Precision + Recall
(5)
MSE: Average squared difference between the estimated values and the actual value.
MSE = MSE x c , y c , a , b , θ = 1 M i = 1 M x c i x ^ c i 2 + y c i y ^ c i 2 + a i a ^ i 2 + b i b ^ i 2 + θ i θ ^ i 2
where M is the number of true positive samples, x c i , y c i , a i , b i , θ i are the predicted values, and x ^ c i , y ^ c i , a ^ i , b ^ i , θ ^ i are the true values.
(6)
Frames Per Second (FPS): the frequency (rate) at which consecutive images (frames) are inferred
(7)
FLOPs: Floating point operations to measure the complexity of the CNN model.
The metrics of precision, recall, AP50, and F1 score are used to evaluate the model’s ability in bubble detection, while MES is used to evaluate the model’s ability to fit elliptical parameters.

3.3. Analysis of Backbone Networks

In this study, we initially assess the efficacy of our approach by testing different backbone network structures to identify the optimal model architecture. We analyzed and compared 11 mainstream YOLO networks, including YOLOv3 [30], YOLOv3-spp [30], YOLOv3-tiny [30], YOLOv4-L [31], YOLOv4-M [31], YOLOv4-S [31], YOLOv5-L [36], YOLOv5-M [36], YOLOv5-S [36], YOLOv7 [32], and YOLOv7-tiny [32]. It should be noted that during the experiment, we only used the network architectures of different YOLO versions as backbone networks. However, some tricks, such as CutMix and Mosaic data augmentation in YOLOv4, were not exploited, and the method used to predict position parameters is different from that used in the original paper.
We utilized a server equipped with an NVIDIA TITAN Xp GPU to train and test the model. During the model training process, the following parameter settings were utilized: input size = 640 × 640 × 3 ; the optimization algorithm was set to Adam; batch size = 32; epochs = 100; and initial learning rate = 1 × 10 3 . Additionally, we introduced weight decay with a factor of 1 × 10 4 to tackle overfitting. The three constant coefficients in the loss function were set to λ l = 0.5 , λ r = 0.5 , and λ c = 0.1 . The IoU threshold for training was set to 0.2.
The model performs multi-head detection on three different feature maps (three branches) of different sizes, which are 80 × 80 , 40 × 40 , and 20 × 20 . Each detection branch has three anchors, resulting in a total of B = 9 3 × 3 anchors. Therefore, the K-means algorithm is used to cluster the parameters in the training set, with a total of nine clusters. The clustering results are (13, 10), (26, 20), (33, 28), (40, 36), (45, 25), (48, 43), (54, 34), (56, 51), and (72, 64). Due to the consistent background generated in the dataset, random flipping is used as the only data augmentation method during training.
Since BubCNN [22] is considered to be the state-of-the-art method for bubble detection, we conducted a comparative analysis between our method and BubCNN. BubCNN was trained for 50 epochs with an initial learning rate of 1 × 10 4 . The learning rate was halved every 10 epochs. During training, 1000 region proposals were evaluated, and the overlap thresholds IoU max and IoU min were set to 0.7 and 0.3, respectively.
Based on the evaluation metrics defined in the previous section, we compared the performance of 11 different backbone network models as shown in Table 1. It can be seen that the precision of 11 different backbone network models is above 0.97, among which the precision of YOLOv5-L backbone network model is the highest, reaching 0.9879. The model with the lowest precision is the one in which the backbone network is YOLOv4-S, and the precision is 0.9750. The precision of these two models differs only by 0.0129. The average precision of the 11 different backbone models is 0.9817, and the variance is 2.12 × 10 5 . Due to the absence of noise interference in the dataset images used in this paper, all models exhibit high precision with minimal differences. The precision of BubCNN is 0.8130. Compared with BubCNN, the precision of YOLOv5-L model is increased by 0.1749, and the precision of YOLOv4-S model is increased by 0.1620.
The difficulty studied in this paper is the detection of bubbles under dense conditions, so recall is a more important metric to focus on. the recall of the model with YOLOv7 as the backbone network is the highest among all models, reaching 0.8725. The recall of the model with YOLOv3-tiny as the backbone model is the lowest, with a value of 0.8310. The recall of the YOLOv7 model differs from that of the YOLOv3-tiny model by only 0.0415. The average recall rate of the 11 different backbone models is 0.8598, and the variance is 1 × 10 4 . Although the precision differences of models with different backbone networks are relatively small, there are still significant differences in the recall rates. The recall rate of BubCNN is only 0.5100, which is increased by 0.3625 when compared to YOLOv7, and increased by 0.3210 when compared to YOLOv3-tiny. This indicates that the method proposed in this paper has good detection performance for bubbles under dense conditions.
The advantage of the AP metric is that it considers a balance between precision and recall at different confidence threshold levels and can be used to compare the performance of different models. The F1 score is a statistical measure used to evaluate the precision of binary classification models, as it considers both precision and recall. Based on the analysis of the AP50 and the F1 score, the best-performing model among all the tested models is the one that uses the YOLOv7 backbone network, with AP50 = 0.9161 and F1 score = 0.9263. YOLOv3 uses Darknet53 as the backbone network. YOLOv4 adds the CSPDarknet53 structure and the Mish activation function on the basis of YOLOv3. YOLOv5 stacks the ConvBNSiLU module, C3 module, and SPPF module to form the backbone, and the YOLOv7 backbone network is composed of the CBS module, ELAN module, and MPConv module. The experimental results indicate that the network structure design of YOLOv7 is more suitable for the detection of dense bubbles. The AP50 result of the YOLOv7 model surpasses BubCNN (0.6440) by 42%, demonstrating the superior performance of the method proposed in this paper for bubble localization.
The MSE metric is an overall evaluation of the model’s fitting performance for two positional parameters x c , y c and three size parameters a , b , θ . From the data in Table 1, the model that has the best fitting performance is the one that uses the YOLOv4-L network, with an MSE error of 3.7273 (in pixels). However, the YOLOv4-L model has average results in terms of precision, recall, AP50, and other metrics. The model with the largest error is the one based on the YOLOv3-tiny structure, with an MSE of 7.2400, due to the shallowest layer of the YOLOv3-tiny model. The average MSE of 11 different backbone models is 4.6765, with a variance of 0.9621. In terms of precision, recall rate, AP50, F1 score, and MSE, the distribution of model performance on the MSE metric is the largest, because the fitting of position and size is evaluated at the pixel level, while other metrics consider predicted results as true positive samples when the IoU between the predicted result and the label is larger than 0.5. The MSE of the YOLOv7 model is 3.8299, which is only 0.1026 higher than the best result, and it is 9.3701 higher than the MSE of BubCNN (13.2). Therefore, the performance of the model based on the YOLOv7 backbone network is excellent as well in terms of parameter fitting.
Taking all evaluation metrics into account, the results in Table 1 demonstrate that the proposed method has good detection performance and parameter fitting precision in dense bubbly flows. Under the premise of considering detection efficiency, it is believed that the model based on the YOLOv7 backbone network has the best overall performance, which is why the analysis of the loss function, IoU, and void fraction in the following sections are based on the YOLOv7 network.

3.4. Analysis of Loss Function and IoU

In object detection, the YOLO algorithm predicts the bounding box for targets through regression. Traditional object detection algorithms use the MSE to predict the coordinates of the center point as well as the length and width of the bounding box, or they predict the upper left and lower right points. With the evolution and iteration of the YOLO algorithm, researchers have used anchor-based methods to estimate the corresponding offset for the position. They use IoU loss to predict the offset of the bounding box relative to the anchor. As the parameters predicted in this study are not bounding boxes, it is necessary to redesign an optimal loss function.
The method proposed in this paper assumes that the bubble is an ellipse rotating counterclockwise around the x-axis. The loss function is composed of three parts: location loss L l o c , regression loss L r e d , and confidence loss L c o n f . The location loss is calculated in the form of CIoU loss.
The effects of loss coefficient λ l , λ r , and λ c for training were investigated. These three parameters were chosen based on empirical methods. Here, we conducted tests using various combinations selected from three values: 0.1, 0.5, and 1.0. The results are given in Table 2. In general, the selection of various combinations has a minor impact on both the location performance and the fitting accuracy of the model. When the values of coefficients λ l and λ r are small, it results in a slight increase in the MSE. This occurs because reducing these two coefficients weakens the constraints on the location and size parameters during the training process. Comparatively, we prefer to improve the location performance of the bubbles. We ultimately selected a parameter combination with the highest AP50 value, which corresponds to λ l = 0.5 , λ r = 0.5 , and λ c = 0.1 .
When calculating IoU, the circumferential horizontal rectangle of the ellipse is taken as the target position. The center coordinates x c , y c , width b w , and height b h of the circumferential horizontal rectangle are calculated by Equations (6) and (7), respectively. Theoretically, it is possible to train the model using only IoU loss through backpropagation to fit the five parameters t x , t y , t a , t b , θ . However, in the experimental process, we found that only fitting the size parameters of the ellipse does not result in optimal performance. Therefore, we experimented with four different methods, including IoU [40], GIoU [41], DIoU [38], and CIoU [38]; the different regression losses are shown in Equations (16)–(18), respectively. The difference between the three different modes lies in whether to include the MSE of the size parameters a , b , θ in the loss function.
M o d e 1 : L r e g = i = 0 S 2 j = 0 B I i j o b j ( a i j a ^ i j ) 2 + ( b i j b ^ i j ) 2 + ( θ i j θ ^ i j ) 2
M o d e 2 : L r e g = i = 0 S 2 j = 0 B I i j o b j ( θ i j θ ^ i j ) 2
M o d e 3 : L r e g = 0
Table 3 presents the experimental results using different IoU and loss functions. The left-hand side represents bubble detection performance, while the right-hand side reflects bubble parameter fitting performance. When using M o d e 1 to calculate the loss function, the four methods of IoU calculation have little impact on precision. The highest model precision is 0.9871 when using CIoU to calculate the loss function and lowest at 0.9859 when using DIoU. The small difference in precision indicates that almost all detected bubbles are true positives. However, the different IoU calculation methods have a certain impact on the model’s recall. Among them, the model using CIoU has the highest recall at 0.8725 and the best AP50 and F19 scores, reaching 0.9161 and 0.9263, respectively. The DIoU model follows, with a recall 2.7 × 10 4 lower than that of the CIoU model, and AP50 and F1 scores decreasing by 2.5 × 10 4 and 2.0 × 10 4 , respectively. The IoU and GIoU methods have little difference in recall, AP50, and F1 score metrics. When all models use CIoU methods, more parameters considered for regression loss ( M o d e 1 ) resulted in better results in terms of precision, recall, AP50, F1 score, and other aspects. Overall, different loss functions have a relatively small difference in bubble detection performance, with slight fluctuations in numerical results but have a significant impact on the fitting of size parameters.
Analysis of the data in the right half of Table 3 shows that the design of regression loss is crucial for the accuracy of parameter fitting. Traditional object detection algorithms only use IoU to predict the bounding box parameters of objects. However, when using IoU loss only to fit the location and size parameters of bubbles in this paper, the fitting performance of the model is the worst, especially for size parameter fitting with significant deviations. Although the CIoU + M o d e 3 method had the largest fitting MSE of location parameters, the difference was only under two, so the performance is considered to be at the same level as the other methods. However, the MSE of parameters a , b are 14.4 and 10.3, higher than the best model (CIoU + M o d e 1 ) by 5.2 and 4.1, respectively. M o d e 2 and M o d e 3 did not calculate the L 2 loss of a , b , so the MSE of these two parameters were larger than those of other models. The CIoU + M o d e 3 model had an MSE of 0.288 for the parameter θ , while the best model (CIoU + M o d e 2 ) had a significantly lower MSE of only 0.0364, which was 7.9 times smaller. The MSEs of other models for the parameter were around 0.04 because they calculated the L 2 loss of θ . Therefore, calculating the L 2 loss of a , b , θ in regression loss can significantly improve the model’s parameter detection accuracy.
Through the experimental analysis in this section, we found that the optimal solution for the loss function is to use CIoU to calculate the location loss and M o d e 1 to calculate the regression loss. Moreover, compared to BubCNN’s performance, the CIoU + M o d e 1 method has shown significant improvements in all performance evaluation metrics, particularly in recall and MSE. In terms of bubble localization, precision increased by 0.1741, recall increased by 0.3625, AP50 increased by 0.2721, and F1 score increased by 0.2995. In terms of parameter fitting, the MSE of parameter x c and parameter y c decreased by 13.8 and 17.4, respectively. The MSE of parameter a and parameter b decreased by 5.7 and 9.7, respectively. MSE of parameter θ decreased by 0.4435, and the average MSE of all five parameters decreased by 9.4.
The design of the IoU loss function in object detection has the potential to accurately predict the bounding box of the target, which comprises four parameters. However, our research focuses on bubble objects that differ from traditional bounding boxes and necessitate the prediction of five parameters. The prediction of parameters x c and y c is equivalent to estimating the center coordinates of the bounding box. Therefore, the IoU loss can be employed. However, since the prediction of parameters a, b, and θ differs from the detection of bounding box width and height, the L 2 loss is employed to enforce the constraint. The optimized loss function offers the following advantages. Firstly, it fulfills the model’s requirement to simultaneously output both localization and size parameters. Secondly, it substantially enhances the accuracy of predicting the bubble’s size parameters, particularly parameter θ . Thirdly, it results in a marginal improvement in the localization performance of the bubbles.

3.5. Bubble Size Distribution Estimation Evaluation

This section evaluates the estimated results of the BSD from two perspectives: the number of bubbles and the size of the bubbles. Figure 6 illustrates the number of bubbles estimated from the test set of 2000 images. The blue curve represents the actual number of bubbles in each sample image, while the yellow curve represents the estimated number of bubbles. For better visualization, minor adjustments were made during the plotting process, including sorting the actual bubble quantities in ascending order. From the figure, it can be observed that when the number of bubbles in each image is less than approximately 60, the model can detect almost all of them. However, as the number of bubbles increases, the deviation between the predicted and actual values gradually becomes larger. The increase in bubble quantity implies a higher void fraction, resulting in more challenging bubble detection, which aligns with our intuitive understanding. The following sections will provide a detailed numerical analysis of precision and recall.
The pixel lengths of the semi-major axis a in the test set were divided into four intervals: 2 < a 10 , 10 < a 20 , 20 < a 30 , and 30 < a 40 . In physical space, every 25 pixels correspond to a length of 1 mm. The estimated results of BSD are shown in Figure 7, where the blue curve represents the actual values and the yellow curve represents the estimated values. Similarly, for the purpose of analysis, we sorted the bubbles in ascending order according to the actual value of a. Therefore, the actual value curves in Figure 7a,d,f,g exhibit a monotonically increasing trend. The analysis of MSE data in Table 4 reveals a clear trend: as the bubble size increases, the fitting errors for parameters a and b also increase. Notably, the MSE of parameter b is relatively smaller in comparison to parameter a. When 2 < a 10 , the MSE values for a and b are only 0.5 and 0.3, respectively. However, when 30 < a 40 , their MSE values increase to 9.2 and 6.5. This is due to the presence of densely distributed bubbles, where larger bubbles are more likely to overlap and occlude other bubbles. The coexistence of large and small bubbles introduces noise and impacts the analysis of the image. Consequently, larger bubble sizes result in larger errors. However, as the size of the bubbles increase, the MSE of parameter θ gradually decreases from 0.0796 to 0.0355. This is because when the bubble size is too small, the rotational effect is not visually significant, making detection difficult. As the bubble size increases, although interference from other bubbles may be present, this noise has a smaller impact on rotation.

3.6. Analysis under Different Void Fractions

Based on the previous experiment, we have explored the best comprehensive performance solution, which uses YOLOv7 as the backbone network and calculates the localization loss and regression loss separately using the CIoU method and Mode1 method. This solution has shown excellent performance in terms of bubble precision, recall, fitting accuracy, and efficiency. In this section, we will explore the detection results of this model under different void fraction conditions.
Figure 8 illustrates the detection results of the best-performing model under different void fractions. The void fraction represents the proportion of the area occupied by bubbles in two-phase flow. The larger the void fraction, the larger the area occupied by bubbles, and the denser the distribution. The void fraction ranges from 0.01 to 0.10 with a step size of 0.01. The images in Figure 8 represent typical pictures at each void fraction. The green dots in each subgraph represent the fitted elliptical bubble center, and the red edge represents the boundary of the ellipse. As evident from the figure, the proposed model accurately detects bubbles and fits elliptical parameters for both sparse and dense bubble distributions.
Table 5 shows the detection performance metrics of the model at different void fractions, with 200 test images at each void fraction. The results of precision, recall, AP50, and F1 score all decrease as the void fraction increases, with recall decreasing the most, from 0.9794 to 0.8120. Precision decreases the least, from 0.9988 to 0.9826, decreasing by 0.0162 in total. The greater decrease in recall indicates an increased rate of false positive detections as the bubbles become denser. As observed from Figure 8, the primary reason for this is that some bubbles are significantly occluded and cannot be detected when the bubbles become denser. The minor decrease in precision is caused by the detection of redundant objects in areas with heavily stacked clusters of bubbles. The multiple edges of tightly packed bubbles in clusters can visually resemble new bubbles, as shown in the top-left region of the bubble cluster in Figure 8h. The slightly decreasing trend of AP50 and F1 score by 0.0808 and 0.0998, respectively, suggests that the model still achieves excellent detection performance in dense bubble distributions.
From the overall perspective, the MSE increases as the void fraction increases. The increases in the MSE of x c , y c , a, b and θ are 2.0, 1.4, 3.3, 2.2, and 0.0304, respectively. Considered within the value ranges of the five parameters (as shown in Figure 9), the changes in the MSE increase relatively slowly. This suggests that the model’s fitting ability is robust and resilient. From Figure 8, it is observed that the reason for the increase in the MSE is due to the possibility of fragmented and overlapped bubbles with dense bubble distributions. The fitting results of these bubbles by the model may exhibit some bias, as seen in the top-right edge of the bubble in Figure 8d and the bottom-right edge of the bubble in Figure 8g.
Figure 9 presents the box plot of MSE errors for the parameter fitting. The figure indicates that the increase in void fraction has a minor impact on the median of the MSE errors for all five parameters, suggesting that the central distribution of fitting errors remains relatively unchanged and positively skewed. However, for parameters x c , y c , and θ , the increase in void fraction results in an increase in the length of the box, and the upper margin shows exponential growth, enhancing the positively skewed distribution. The upper edge of parameters a and b slightly increases with the increase in void fraction, and its error distribution is more stable in comparison to the other three parameters. The appearance of a higher upper edge indicates the occurrence of missed detections (false positives) in rare samples. Figure 9 further suggests that the fitting results for most bubbles are minimally affected by the increase in void fraction, but the main impact is on the occurrence of anomalous detection results.
Overall, the model proposed in this paper exhibits high accuracy and robustness in detecting dense bubbles and fitting parameters.

4. Discussions

The traditional image-processing-based bubble detection method can achieve good detection performance under specific experimental conditions. However, changes in the usage scenario often lead to a decrease in algorithm performance or even make it unsuitable for use. BubCNN was the first to apply the deep-learning-based object detection network Faster-RCNN to the field of bubble detection and combine it with a shape regression CNN network for bubble shape parameter detection to create a two-stage bubble detector. Following this path of object detection, this paper designs an end-to-end bubble detection network based on the YOLO network to achieve more efficient bubble location and parameter detection. Through experimental comparative analysis, our proposed method has the following advantages: Firstly, for densely distributed bubbles, the proposed model greatly improves the precision and recall over existing methods. The model has better overall bubble positioning performance and has smaller fitting errors for parameters. Secondly, the improved bubble detection network proposed, which is based on the YOLO method, is a one-stage network with higher detection efficiency and real-time performance. The fastest model achieves an FPS of 255. Lastly, the open-source BubGAN dataset was used for training the data. The diverse shapes of bubbles in the dataset enhance the model’s generalization ability and extend its applicability to a wider range.
Nevertheless, the proposed method also has some limitations. Firstly, in highly dense bubble distributions, the proposed method may produce false positive missed detections and erroneous parameter detection results for bubbles that are incomplete or fragmented at the edges of the image. In addition, although the assumption that the bubble contour is usually elliptical is valid, it becomes inaccurate when dealing with higher Eotvos and Reynolds numbers. Therefore, deep-learning-based instance segmentation methods can be explored as potential solutions to address these two issues. Bubbles are segmented with pixel-wise precision, and morphological parameters of each small bubble can be extracted as needed. Finally, it should be noted that this research is based on 2D bubble images, while 3D bubbles may lose some size features after being sampled by 2D cameras. The stereological correction was not taken into consideration to determine bubble size, which may result in certain errors.

5. Conclusions

To address the problem of difficult detection of bubbles in dense bubbly flow, this paper proposes a novel BSD detection method based on deep learning. The main contributions of this paper are as follows:
(1)
An end-to-end BSD detection method based on YOLO is proposed, using a multi-head detection method to detect different-sized bubbles from three scales, and detecting dense objects by using a dense detection approach. The model also adds an ellipse fitting output for the morphological parameters of the bubbles, achieving a synchronous output of their position and position parameters.
(2)
The loss function of dense bubble detection is optimized using CIoU of bubble objects and L 2 constraints of elliptical parameters to improve the accuracy of model parameter fitting. The precision, recall, and AP50 of the model are 0.9871, 0.8725, and 0.9161 respectively, and the MSE of the parameters is 3.8299.
Furthermore, pixel-level detection of bubbles will be researched in further work by using deep-learning-based instance segmentation techniques.

Author Contributions

Conceptualization, W.H. and S.Z.; methodology, M.C., C.Z. and S.Z.; software, M.C., C.Z. and W.Y.; validation, C.Z. and S.Z.; formal analysis, M.C., C.Z. and W.Y.; investigation, S.Z. and W.H.; resources, W.H.; data curation, C.Z. and W.Y.; writing—original draft preparation, M.C. and S.Z.; writing—review and editing, C.Z., W.Y. and W.H.; visualization, M.C. and S.Z.; supervision, S.Z. and W.H.; project administration, W.H.; funding acquisition, W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (2022YFB4201603).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to express our appreciation to Mingxuan Zou and Junpeng Mi for their insightful guidance in methodology, and to Shuai Zhang and Debin Liu for their invaluable assistance in code implementation. We would also like to acknowledge the computing device support from Zhejiang SUPCON Technology Co., Ltd.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Peng, C.; Liu, Y.; Gui, W.; Tang, Z.; Chen, Q. Bubble image segmentation based on a novel watershed algorithm with an optimized mark and edge constraint. IEEE Trans. Instrum. Meas. 2021, 71, 5005110. [Google Scholar] [CrossRef]
  2. Chen, M.; Liu, H.; Zhang, S.; Liu, Z.; Mi, J.; Huang, W.; Li, D. Spirits quality classification based on machine vision technology and expert knowledge. Meas. Sci. Technol. 2023, 34, 055405. [Google Scholar] [CrossRef]
  3. Wang, D.; Dai, W.; Tang, D.; Liang, Y.; Ouyang, J.; Wang, H.; Peng, Y. Deep learning approach for bubble segmentation from hysteroscopic images. Med. Biol. Eng. Comput. 2022, 60, 1613–1626. [Google Scholar] [CrossRef] [PubMed]
  4. Barigou, M.; Greaves, M. A capillary suction prove for bubble size measurement. Meas. Sci. Technol. 1991, 2, 318. [Google Scholar] [CrossRef]
  5. Liu, T.; Bankoff, S. Structure of air-water bubbly flow in a vertical pipe—II. Void fraction, bubble velocity and bubble size distribution. Int. J. Heat Mass Transf. 1993, 36, 1061–1072. [Google Scholar] [CrossRef]
  6. Saberi, S.; Shakourzadeh, K.; Bastoul, D.; Militzer, J. Bubble size and velocity measurement in gas—liquid systems: Application of fiber optic technique to pilot plant scale. Can. J. Chem. Eng. 1995, 73, 253–257. [Google Scholar] [CrossRef]
  7. Prasser, H.M. Novel experimental measuring techniques required to provide data for CFD validation. Nucl. Eng. Des. 2008, 238, 744–770. [Google Scholar] [CrossRef] [Green Version]
  8. Laakkonen, M.; Moilanen, P.; Miettinen, T.; Saari, K.; Honkanen, M.; Saarenrinne, P.; Aittamaa, J. Local bubble size distributions in agitated vessel: Comparison of three experimental techniques. Chem. Eng. Res. Des. 2005, 83, 50–58. [Google Scholar] [CrossRef]
  9. Garcia-Magarino, A.; Sor, S.; Bardera, R.; Munoz-Campillejo, J. Interferometric laser imaging for droplet sizing method for long range measurements. Measurement 2021, 168, 108418. [Google Scholar] [CrossRef]
  10. Honkanen, M.; Saarenrinne, P.; Stoor, T.; Niinimäki, J. Recognition of highly overlapping ellipse-like bubble images. Meas. Sci. Technol. 2005, 16, 1760. [Google Scholar] [CrossRef]
  11. Talbot, H. Elliptical distance transforms and applications. In Proceedings of the International Conference on Discrete Geometry for Computer Imagery, Szeged, Hungary, 25–27 October 2006; pp. 320–330. [Google Scholar]
  12. Hukkanen, J.; Hategan, A.; Sabo, E.; Tabus, I. Segmentation of cell nuclei from histological images by ellipse fitting. In Proceedings of the 18th European Signal Processing Conference, Aalborg, Denmark, 23–27 August 2010; pp. 1219–1223. [Google Scholar]
  13. Zafari, S.; Eerola, T.; Sampo, J.; Kälviäinen, H.; Haario, H. Segmentation of overlapping elliptical objects in silhouette images. IEEE Trans. Image Process. 2015, 24, 5942–5952. [Google Scholar] [CrossRef]
  14. Lau, Y.; Deen, N.; Kuipers, J. Development of an image measurement technique for size distribution in dense bubbly flows. Chem. Eng. Sci. 2013, 94, 20–29. [Google Scholar] [CrossRef]
  15. Zhang, W.H.; Jiang, X.; Liu, Y.M. A method for recognizing overlapping elliptical bubbles in bubble image. Pattern Recognit. Lett. 2012, 33, 1543–1548. [Google Scholar] [CrossRef]
  16. Zafari, S.; Eerola, T.; Sampo, J.; Kälviäinen, H.; Haario, H. Segmentation of partially overlapping convex objects using branch and bound algorithm. In Proceedings of the Asian Conference on Computer Vision, Taiwan, China, 20–24 November 2017; pp. 76–90. [Google Scholar]
  17. Fu, Y.; Aldrich, C. Flotation froth image analysis by use of a dynamic feature extraction algorithm. IFAC PapersOnLine 2016, 49, 84–89. [Google Scholar] [CrossRef]
  18. Taboada, B.; Vega-Alvarado, L.; Córdova-Aguilar, M.; Galindo, E.; Corkidi, G. Semi-automatic image analysis methodology for the segmentation of bubbles and drops in complex dispersions occurring in bioreactors. Exp. Fluids 2006, 41, 383–392. [Google Scholar] [CrossRef]
  19. Ilonen, J.; Juránek, R.; Eerola, T.; Lensu, L.; Dubská, M.; Zemčík, P.; Kälviäinen, H. Comparison of bubble detectors and size distribution estimators. Pattern Recognit. Lett. 2018, 101, 60–66. [Google Scholar] [CrossRef]
  20. Poletaev, I.; Pervunin, K.; Tokarev, M. Artificial neural network for bubbles pattern recognition on the images. J. Phys. Conf. Ser. 2016, 754, 072002. [Google Scholar] [CrossRef]
  21. Poletaev, I.; Tokarev, M.P.; Pervunin, K.S. Bubble patterns recognition using neural networks: Application to the analysis of a two-phase bubbly jet. Int. J. Multiph. Flow 2020, 126, 103194. [Google Scholar] [CrossRef]
  22. Haas, T.; Schubert, C.; Eickhoff, M.; Pfeifer, H. BubCNN: Bubble detection using Faster RCNN and shape regression network. Chem. Eng. Sci. 2020, 216, 115467. [Google Scholar] [CrossRef]
  23. Cerqueira, R.F.; Paladino, E.E. Development of a deep learning-based image processing technique for bubble pattern recognition and shape reconstruction in dense bubbly flows. Chem. Eng. Sci. 2021, 230, 116163. [Google Scholar] [CrossRef]
  24. Hernandez, J. A technique for the direct measurement of bubble size distributions in industrial flotation cells. In Proceedings of the 34th Annual Meeting of the Canadian Mineral Processors, Ottawa, ON, Canada, 22–24 January 2002. [Google Scholar]
  25. Ma, Y.; Yan, G.; Scheuermann, A.; Bringemeier, D.; Kong, X.Z.; Li, L. Size distribution measurement for densely binding bubbles via image analysis. Exp. Fluids 2014, 55, 1860. [Google Scholar] [CrossRef]
  26. Azgomi, F.; Gomez, C.; Finch, J. Correspondence of gas holdup and bubble size in presence of different frothers. Int. J. Miner. Process. 2007, 83, 1–11. [Google Scholar] [CrossRef]
  27. Vinnett, L.; Urriola, B.; Orellana, F.; Guajardo, C.; Esteban, A. Reducing the presence of clusters in bubble size measurements for gas dispersion characterizations. Minerals 2022, 12, 1148. [Google Scholar] [CrossRef]
  28. Liu, J.; Gao, Q.; Tang, Z.; Xie, Y.; Gui, W.; Ma, T.; Niyoyita, J.P. Online monitoring of flotation froth bubble-size distributions via multiscale deblurring and multistage jumping feature-fused full convolutional networks. IEEE Trans. Instrum. Meas. 2020, 69, 9618–9633. [Google Scholar] [CrossRef]
  29. Fu, Y.; Liu, Y. BubGAN: Bubble generative adversarial networks for synthesizing realistic bubbly flow images. Chem. Eng. Sci. 2019, 204, 35–47. [Google Scholar] [CrossRef] [Green Version]
  30. Farhadi, A.; Redmon, J. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  31. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  32. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2023, arXiv:2207.02696. [Google Scholar]
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  35. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  36. Jocher, G. YOLOv5 by Ultralytics. 2020. Available online: https://zenodo.org/record/4418161 (accessed on 15 November 2022).
  37. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  38. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000.
  39. Vinnett, L.; Yianatos, J.; Arismendi, L.; Waters, K. Assessment of two automated image processing methods to estimate bubble size in industrial flotation machines. Miner. Eng. 2020, 159, 106636. [Google Scholar] [CrossRef]
  40. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 516–520. [Google Scholar]
  41. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE conference on computer vision and pattern recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
Figure 1. End-to-end BSD detection workflow.
Figure 1. End-to-end BSD detection workflow.
Sensors 23 06582 g001
Figure 2. The proposed one-stage BSD detector architecture based on YOLO. The orange and blue blocks represent the backbone module and the neck module, respectively.
Figure 2. The proposed one-stage BSD detector architecture based on YOLO. The orange and blue blocks represent the backbone module and the neck module, respectively.
Sensors 23 06582 g002
Figure 3. Detector output illustration.
Figure 3. Detector output illustration.
Sensors 23 06582 g003
Figure 4. Ellipse fitting with dimension priors.
Figure 4. Ellipse fitting with dimension priors.
Sensors 23 06582 g004
Figure 5. The distribution of dataset parameters.
Figure 5. The distribution of dataset parameters.
Sensors 23 06582 g005
Figure 6. Comparison of the estimated and actual number of bubbles.
Figure 6. Comparison of the estimated and actual number of bubbles.
Sensors 23 06582 g006
Figure 7. Comparison of the estimated and actual bubble size. The sub-figures in the first, second, and third columns display the comparison results between the estimated and actual values of a, b, and θ under different ranges of the semi-major axis (a). (ac): 2 < a 10 ; (df): 10 < a 20 ; (gi): 20 < a 30 ; (jl): 30 < a 40 .
Figure 7. Comparison of the estimated and actual bubble size. The sub-figures in the first, second, and third columns display the comparison results between the estimated and actual values of a, b, and θ under different ranges of the semi-major axis (a). (ac): 2 < a 10 ; (df): 10 < a 20 ; (gi): 20 < a 30 ; (jl): 30 < a 40 .
Sensors 23 06582 g007
Figure 8. Detection results under different void fractions.
Figure 8. Detection results under different void fractions.
Sensors 23 06582 g008
Figure 9. Mean square error results of different void fractions.
Figure 9. Mean square error results of different void fractions.
Sensors 23 06582 g009
Table 1. Detector performance of different backbone networks.
Table 1. Detector performance of different backbone networks.
BackbonePrecisionRecallAP50F1 ScoreMSEFLOPsParametersFPS
YOLOv70.98710.87250.91610.92633.8299103.2G36,497,95455
YOLOv7-tiny0.98430.86140.91520.91874.679613.0G6,015,714144
YOLOv5-L0.98790.8640.9150.92183.74107.7G46,124,43351
YOLOv5-M0.98260.8630.91310.91894.623947.9G20,865,05776
YOLOv5-S0.98310.8540.91370.9144.546115.8G7,020,913135
YOLOv4-L0.9840.8620.90670.9193.7273119.1G52,496,68946
YOLOv4-M0.9860.8650.90630.92154.150.3G24,357,84969
YOLOv4-S0.9750.8610.9030.91454.864520.6G9,118,721117
YOLOv30.9760.85710.91160.91274.39154.6G61,513,58541
YOLOv3-spp0.9770.8670.91320.91875.7155.4G62,562,67340
YOLOv3-tiny0.9760.8310.9030.89777.2412.9G8,673,622250
BubCNN0.8130.510.6440.626813.2--2
Table 2. Detector performance with different coefficients for the loss function.
Table 2. Detector performance with different coefficients for the loss function.
λ 1 λ r λ c PrecisionRecallAP50F1 ScoreMSE
0.50.50.50.98480.87380.90630.92603.9323
1.00.50.50.98530.87380.90670.92623.7166
0.51.00.50.98700.87130.90620.92553.8856
0.50.51.00.98420.87230.90580.92493.9255
0.10.50.50.98320.87380.90660.92524.0088
0.50.10.50.98590.87180.91350.92544.2846
0.50.50.10.98710.87250.91610.92633.8299
Table 3. Analysis of loss function and IOU.
Table 3. Analysis of loss function and IOU.
Loss FunctionPrecisionRecallAP50F1 ScoreMSE
x c y c ab θ Total
IoU + M o d e 1 0.98610.86540.90870.92182.41.811.37.40.04434.5852
GIoU + M o d e 1 0.98630.86970.90930.92432.92.112.99.80.04495.5473
DIoU + M o d e 1 0.98590.86990.91360.92432.31.710.17.20.04074.2846
CIoU + M o d e 1 0.98710.87250.91610.92632.11.69.26.20.03853.8299
CIoU + M o d e 2 0.98340.8670.91030.92152.51.911.911.70.03645.6338
CIoU + M o d e 3 0.97920.8610.91160.91633.32.514.410.30.28886.1692
BubCNN0.8130.510.6440.626815.91914.915.90.482013.2364
Table 4. MSE of bubbles with different semi-major axis sizes.
Table 4. MSE of bubbles with different semi-major axis sizes.
Semi-Major Axis Range 2 < a 10 10 < a 20 20 < a 30 30 < a 40
MSE of a0.51.53.79.2
MSE of b0.31.02.66.5
MSE of θ 0.07960.04150.03480.0355
Table 5. Test results under different void fractions.
Table 5. Test results under different void fractions.
Void FractionPrecisionRecallAP50F1 ScoreMSE
x c y c ab θ Total
0.010.99880.97940.98560.9890.40.46.34.40.01652.2949
0.020.99560.96020.97650.977610.874.70.02342.7038
0.030.99620.94560.97040.97031.4185.50.02533.1874
0.040.99250.93430.96370.96251.81.495.90.02883.6188
0.050.9910.91160.95310.94962.21.69.76.40.03514.0073
0.060.99060.89050.94410.93792.41.79.56.30.03714.0003
0.070.98530.87470.9360.92672.21.79.26.30.0413.8744
0.080.98480.85710.92790.91652.21.69.46.40.04213.9475
0.090.98120.83540.91620.90242.41.89.56.40.04424.0237
0.10.98260.8120.90480.88922.41.89.66.60.04694.1023
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, M.; Zhang, C.; Yang, W.; Zhang, S.; Huang, W. End-to-End Bubble Size Distribution Detection Technique in Dense Bubbly Flows Based on You Only Look Once Architecture. Sensors 2023, 23, 6582. https://doi.org/10.3390/s23146582

AMA Style

Chen M, Zhang C, Yang W, Zhang S, Huang W. End-to-End Bubble Size Distribution Detection Technique in Dense Bubbly Flows Based on You Only Look Once Architecture. Sensors. 2023; 23(14):6582. https://doi.org/10.3390/s23146582

Chicago/Turabian Style

Chen, Mengchi, Cheng Zhang, Wen Yang, Suyi Zhang, and Wenjun Huang. 2023. "End-to-End Bubble Size Distribution Detection Technique in Dense Bubbly Flows Based on You Only Look Once Architecture" Sensors 23, no. 14: 6582. https://doi.org/10.3390/s23146582

APA Style

Chen, M., Zhang, C., Yang, W., Zhang, S., & Huang, W. (2023). End-to-End Bubble Size Distribution Detection Technique in Dense Bubbly Flows Based on You Only Look Once Architecture. Sensors, 23(14), 6582. https://doi.org/10.3390/s23146582

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop