Next Article in Journal
Characteristics and Metallogenic Significance of Fe-Mn Carbonate Minerals in the Erdaokan Ag Deposit, Heilongjiang Province, Northeast China: Constraints from Sm-Nd Geochronology and Trace Elements
Previous Article in Journal
Editorial for the Special Issue: “Valuable Metals Recovery by Mineral Processing and Hydrometallurgy”
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identification of Rock Fragments after Blasting by Using Deep Learning-Based Segment Anything Model

1
School of Resources and Safety Engineering, Central South University, Changsha 410083, China
2
Department of Civil Engineering, Monash University, Melbourne, VIC 3800, Australia
3
CINF Engineering Co., Ltd., Changsha 410001, China
*
Author to whom correspondence should be addressed.
Minerals 2024, 14(7), 654; https://doi.org/10.3390/min14070654
Submission received: 23 May 2024 / Revised: 22 June 2024 / Accepted: 24 June 2024 / Published: 26 June 2024
(This article belongs to the Special Issue Sustainable Mining: Advancements, Challenges and Future Directions)

Abstract

:
Rock fragmentation is an important evaluation indicator for field blasting operations. This paper applies a deep learning-based method, the Segment Anything Model (SAM), to automatically segment rock fragments. To review the SAM’s segmentation performance, 83 images of rock fragment collected from the mine site were used as the test dataset. Pixel-level accuracy (PA), intersection over union (IOU), and dice coefficient (Dice) were employed to evaluate the model pixel-level segmentation performance. The results showed that the SAM exhibited excellent segmentation performance on the test data (PA = 94.5%, IOU = 94.4%, Dice = 95.4%). The coefficient of determination (R2) values for the 50% and 80% passing sizes (X50 and X80) were 0.970 and 0.991, respectively, which demonstrated that the SAM could achieve high precision measurement of rock fragmentation. Additionally, the effectiveness of the SAM was further evaluated by comparing it to commercial software, and the generalizability of the SAM was verified on two other datasets. The findings revealed that the SAM not only outperformed the Split-Desktop V 4.0 on the test dataset but also achieved comparable accuracy to previous studies on the two other datasets. The SAM could be regarded as a useful tool to provide fast and accurate feedback for field blasting.

1. Introduction

Blasting technology is widely used in mining engineering for rock breaking. The size of rock fragments after blasting directly affects subsequent operations, such as shoveling, transportation, and crushing [1,2,3,4]. Maintaining appropriate rock fragment sizes can reduce blasting costs and enhance mining productivity. More importantly, it mitigates the need for a secondary blast, thereby effectively ensuring the safety of operators. The accurate measurement of a rock fragment size is crucial for achieving high efficiency, cost effectiveness, and safety in mining operations.
As an earlier attempt, empirical models were widely accepted for rock fragment prediction, and the most well-known empirical method is the Kuz–Ram model [5]. After that, several modified methods, such as the crush zone model (CZM) [6], two-component model (TCM) [7], Kuznetsov–Cunningham–Ouchterlony model (KCO) [8], and Kou–Rustan model [9], were proposed. However, these empirical models are presented in different mathematic formulas and are difficult to generalize to other datasets. An extensive review of previous studies, including WipFrag [10], Split Desktop [11], FragScan [12], Fragalyst [13], PowerSieve [14], and GoldSize [15], indicates that the image analysis methods provide more simplicity, efficiency, and accuracy compared to traditional empirical approaches. Because of this, many scholars have developed various digital image-based methods for rock fragment segmentation. For example, Amankwah and Aldrich [16] developed a modified watershed method for rock fragment segmentation. The proposed method outperformed the original watershed method in finer particles’ prediction. Sereshki et al. [17] achieved precise rock fragmentation prediction in laboratory scale by integrating filter algorithms, image binarization, edge detection, and image morphology methods. The obtained results were proven to be more accurate than Split Desktop and GoldSize software. Guo et al. [18] created an improved watershed algorithm for rock fragment segmentation using a novel seed point marking strategy. The obtained results were comparable to manual measurements. Yang et al. [19] proposed a superpixel image segmentation method combining the wavelet transform and nonlocal multiscale fractional differentiation method, achieving over 90% accuracy in particle size distribution prediction. Traditional image segmentation methods often involve cumbersome steps to achieve satisfactory results on specific data. It is challenging for them to deal with images with different brightness, shading, and surface texture characteristics. Therefore, it cannot be widely applied and promoted in engineering [19,20,21].
In recent years, deep learning-based convolutional neural networks (CNNs) have demonstrated remarkable performance across various domains of computer vision, achieving human-comparable accuracy in tasks such as image classification [22,23,24,25], object detection [26,27,28,29], and image segmentation [30,31,32]. Specifically in rock fragment segmentation, CNN-based methods have shown significant advantages over traditional image segmentation methods [21,33,34,35,36]. Guo et al. [21] proposed a network for rock fragment segmentation based on multiple CNN structures. A total of 14,628 labeled image patches, extracted from the entire blast rock pile images using data augmentation methods, were used for model development. The model performed quite well on two test images, achieving a higher R2 value of 0.9989. In another study, Bamford et al. [37] collected 61,853 images of 512 × 512 size from 3301 sets of images using data augmentation methods and proposed a deep neural network (DNN) model; the size prediction error of the trained model on coarse blocks (X50, X80, Xmax) was within ±25%, with half of the results being within ±10%. Wang et al. [20] created a multi-task CNN structure for ore segmentation, which included a boundary detection subnetwork and a mask segmentation network. The proposed model achieved robust results. Additionally, Li et al. [38] used the state-of-the-art edge detection model DexiNed network [39] and image morphology techniques to measure particle size on the conveyor belt in an aggregate crushing plant. The proposed method demonstrated its excellent capability to recognize rock edge features and successfully separated coarse fragments from fine contiguous areas. Among CNN architectures, the U-Net proposed by Ronneberger et al. [30] has become the most practical method for rock fragment segmentation due to its remarkable performance. Various extended versions have been proposed and applied in rock fragmentation prediction [40,41,42,43]. Despite their successes, these deep learning approaches heavily rely on large-scale datasets, which often necessitate tedious, time-consuming, and costly data preparation and manual labeling processes. To resolve these defects, this paper adopts the Segment Anything Model (SAM) [44], a deep learning-based image segmentation fundamental model, for automatic rock fragments’ segmentation. Unlike other deep learning models that require numerous labeled datasets for training, the SAM achieves zero-shot segmentation of rock fragments without the need for pre-prepared training data, significantly reducing manual labor costs and enhancing efficiency.
The main contents of this paper are as follows: (1) an overview of the SAM; (2) test data preparation including image acquisition, image quality analysis, and classification; (3) evaluation of the SAM’s zero-shot segmentation performance on test dataset; (4) analysis and discussion of the effectiveness and generalizability of the SAM; and (5) conclusions and perspectives.

2. Segment Anything Model

Pretrained fundamental models such as BERT [45] and GPT series models [46,47] have shown powerful application value in recent years. These large models stemmed from the emergence of big data and the advancements in computing technology. By utilizing the pretrained knowledge of these fundamental models along with appropriate prompts, they can achieve promising results on new tasks with minimal costs. Like large models in the natural language processing (NLP) field, a fundamental image segmentation model, the SAM, was first proposed in the computer vision field [44].
Most deep learning-based image segmentation models are designed for specific tasks (e.g., image semantic segmentation or instance segmentation) and are trained with the specific dataset. However, when the task changes, new datasets need to be collected and labeled to retrain or fine-tune the model to ensure good model performance. To address this limitation, the SAM, a fully automated pixel-level image segmentation fundamental model, was proposed. It was trained on more than 10 billion masks in 11 million images. The SAM achieved outstanding zero-shot segmentation performance on different image tasks without human intervention and provided an important and generalized basic framework for image segmentation within the field of computer vision. Figure 1 shows the zero-shot segmentation capabilities of the SAM on a large dataset.
The SAM consists of three modules: an image encoder, a prompt encoder, and a mask decoder. The image encoder is responsible for understanding and processing the input data, the prompt encoder aims to address the promptable information, and the mask decoder is responsible for generating the valid mask. The image encoder in the SAM employs a vision transformer (ViT-H/16) [48] pretrained using masked autoencoders (MAE) [49]. The ViT-H/16 has been adapted to handle high-resolution inputs with 14 × 14 windowed attention and four equally spaced global attention blocks. This component consists of approximately 632 million parameters and includes 32 transformer layers. The prompt encoder processes various types of prompts, categorized into sparse and dense prompts. Sparse prompts include points, represented as the sum of a positional encoding and a learned embedding indicating foreground or background, and boxes, represented by an embedding pair that includes the positional encoding of the top-left and bottom-right corners summed with learned embeddings representing these corners. For text prompts, the prompt encoder utilizes the text encoder from Contrastive Language-Image Pretraining (CLIP) to embed free-form text. Dense prompts, such as masks, are input at a 4× lower resolution than the input image and further downscaled using two 2 × 2, stride-2 convolutions followed by a 1 × 1 convolution, all separated by GELU activations and layer normalization. The mask decoder integrates the image embedding and prompt embeddings to predict the output mask. It employs a modified transformer decoder block with prompt self-attention and cross-attention (prompt-to-image and image-to-prompt) followed by a dynamic mask prediction head. This head uses a multi-layer perceptron (MLP) to map the output token to a dynamic linear classifier, computing the mask foreground probability at each image location. The mask decoder consists of two transformer decoder layers, each incorporating self-attention and cross-attention mechanisms. The entire SAM model comprises approximately 1 billion parameters, distributed across the image encoder, prompt encoder, and mask decoder. The overall structure of the SAM is shown in Figure 2. More detailed information can be found in the original paper [44] and the following website: https://github.com/facebookresearch/segment-anything (accessed on 5 April 2023).
The dataset used to train the SAM was obtained in three phases. First, annotators manually modified masks produced by the pretrained SAM while simultaneously labeling images. Subsequently, the SAM was retrained using these modified datasets. The retrained SAM then automatically generated masks for the images, and the low-confidence parts were then corrected manually, thereby increasing the variety of objects. After that, the SAM was retrained again. Finally, the labeling operation was completed entirely by the SAM, yielding a training dataset consisting of more than 10 billion masks in 11 million images.
The SAM method has been widely used in several fields, including object detection and tracking [50,51], image processing [52,53,54], and medical image analysis [55,56,57]. However, the application of the SAM in geotechnical and mining engineering, particularly in rock fragment segmentation, has not been reported. Consequently, this study comprehensively analyzed the zero-shot performance of the SAM for rock fragment segmentation.

3. Dataset

3.1. Image Acquisition

In this paper, 83 sets of images were collected from the mine site. The dataset consisted of two parts: one set of images was collected from underground and the other part was captured on the surface. A red pen and flashlight were used as size calibrators throughout the photography process, as shown in Figure 3. Notably, to ensure high-quality image acquisition conducive to accurate segmentation and analysis, specific requirements were followed: (1) Shooting Distance: The camera was positioned at a consistent distance of one meter from the rock fragments, ensuring the capture of clear and detailed images without distortion as much as possible. (2) Shooting Angle: Photographs were taken from a top-down, perpendicular angle to the rock surface whenever possible to minimize perspective distortion and accurately capture the contours of the rock fragments, thereby reducing edge deformations. (3) Lighting Conditions: Adequate and uniform lighting was employed to prevent shadows and highlights that could obscure details. Underground images were supplemented with compensating lamps, while surface images utilized natural light.
The filed images showed a diverse distribution of rock size, including fine particles, coarse grains, and a mixture of both. Additionally, the surfaces of the rock blocks obtained from underground often contained more dirt and dust, resulting in more blurred contour lines. Figure 3 shows some examples of the collected images.

3.2. Image Quality Evaluation

It is well known that image quality is closely related to the segmentation results of image analysis methods. Therefore, in this paper, we first quantitively assessed the image quality based on image sharpness and blurriness. The higher the sharpness value was, the better the image quality was. For blurriness, the smaller the value was, the blurrier the image was.
As shown in Figure 4, the sharpness values ranged from 6 to 26 and the blurriness ranged from 100 to 1500, indicating that the image dataset used in this study presented good diversity. In addition, the calculation results of the two metrics showed a similar distribution, which demonstrates that the selected image quality evaluation metrics were reasonable.

3.3. Data Categorization

To evaluate the performance of the SAM, all the images were grouped based on the obtained results of sharpness and blurriness. For this purpose, the K-means method was used for clustering and the sum of squared errors (SSE) was adopted as the loss function. The clustering loss results are shown in Figure 5. In Figure 5, the left y-axis indicates the loss of sharpness while the right y-axis represents the results of blurriness. It could be summarized that the cluster loss obtained based on the sharpness metric was the smallest, and the best K value for clustering could be determined based on the elbow method (K = 2).
As shown in Figure 6, both the sharpness and blurriness results were well categorized into two classes based on the clustering labels of the sharpness metric. This indicates that the selected sharpness metric in this paper was a better benchmark for rock image quality assessment. Then, all the testing images were ultimately divided into two groups, ranked from high-quality images to poor-quality images, with quantities of 53 and 30, respectively. In addition, no training data needed to be prepared for the SAM; therefore, all captured images were considered as test data.

4. Evaluation of the SAM

As mentioned above, all test data were categorized into two groups: high-quality images and poor-quality images. Then, the ground truth for all test images was created through manual annotations. Based on these well-labeled images, the performance of the SAM in rock fragment segmentation was analyzed in detail in terms of pixel-level segmentation accuracy and size distribution results.

4.1. Performance of the SAM on High-Quality Images

(1)
Pixel-level segmentation results
High-quality images, as defined before, were images with a sharpness value greater than 14 and less than 27. Figure 7 shows the segmentation results of four high-quality test images. As can be seen from Figure 7, different colors indicate different rock blocks, the rock block masks generated by the SAM were consistent with the manual labeling results for both large and small blocks.
The widely used image segmentation evaluation metrics, such as pixel-level accuracy ( P A ), mean intersection over union ( m I O U ), and D i c e   c o e f f i c i e n t , were utilized to quantitatively evaluate the pixel-level segmentation performance of the SAM, and the formulas were as follows:
P A = i = 1 n p i i i = 1 n j = 1 n p i j
m I O U = 1 n + 1 i = 1 n p i i j = 1 n p i j + j = 1 n p j i p i i
D i c e   C o e f f i c e n t = 2 × X Y X + Y
where n is the number of pixel classes in the labeled image (only the fragments and the background were considered in this study, so n = 2 ), p i i denotes the number of pixel points for which i is predicted to be i, p i j represents the number of pixel points for which i is predicted to be j, and p j i reflects the quantity of pixel points for which j is predicted to be i. The Dice coefficient is defined as two times the intersection of the predicted segmentation (X) and the true segmentation (Y) divided by the sizes of X and Y.
Table 1 shows the evaluation results of the SAM on all high-quality images. The scores of three evaluation indicators were 0.955, 0.953, and 0.960, respectively, indicating that the SAM exhibited excellent pixel-level segmentation capabilities when applied to high-quality data.
(2)
Block size distribution results
Additionally, the size distribution of the rock blocks was calculated. First, each segmented mask obtained by the SAM was calibrated using a size calibrator (the red marker pen was 13.7 cm × 1.8 cm in length and diameter, and the flashlight was 16 cm × 2.6 cm). Then, the fragment size was further calculated using the equivalent circle diameter as shown in Equation (4) [42]. Finally, the real-world size of the segmented blocks was obtained. The size distributions of the segmented blocks in the four test images were shown in Figure 8.
D = 2 × S π
where S indicates the segmentation area and D is the diameter of the equivalent circle.
Based on the obtained block size distribution results, the accumulated passing percentage curve could be calculated using the Rosin–Rammler function [5], as follows:
R x = 1 e ( X X c ) n
where R x represents the proportion of material that passes through the screen, X denotes the screen size in cm, X c is the characteristic size that 63.21% of the particles pass (cm), and n is the uniformity index.
According to Equation (5), the characteristic size X c could be expressed as:
X c = X l n ( 1 R ( x ) ) n
Furthermore, as mentioned in [5], when X   =   X m , Equation (6) could also be rewritten as follows:
X c = X m 0.693 n
where X m is the mean fragment size and the value of R ( x ) equals 0.5 when X   =   X m .
Then, Equation (5) could be updated as follows:
R x = 1 e 0.693 ( X X m ) n
The value of parameter n could be calculated based on the well-established formulas in the Kuz–Ram model [5]. However, another calculation method was also recommended [58,59]:
n = 0.842 l n k 80 l n k 50
where k 80 and k 50 are the sieve sizes at 50% and 80% particle passage (cm), respectively.
Figure 9 shows the comparison results of the accumulated passing percentage obtained from the SAM and manual segmentation results on the four testing samples. The dotted lines indicate the SAM method, and the solid line represents the manual segmentation result. It can be seen that the percentage passing curves of the SAM method were very close to the manual segmentation results on the four testing images. Furthermore, the values of 10 characteristic sizes ( X 10 , X 20 ,   , X 100 ) of the four test images obtained by the SAM were compared with the results obtained from manual segmentation, as shown in Table 2. The mean absolute error (MAE), root mean square error (RMSE), and the coefficient of determination (R2) were used to quantitatively assess the SAM performance. The average results for the three evaluation indicators on the four test images were 0.04, 1.25, and 0.98, respectively.
As mentioned in Section 3.3, there were a total of 53 sets of high-quality images. Therefore, the block size distributions for all high-quality samples obtained by the SAM and the manual segmentation were summarized, as shown in Figure 10. The results of characteristic sizes X 10 , X 20 , X 30 , X 40 , and X 50 obtained by the SAM were slightly different compared to the manual segmentation results. The differences were primarily because a few images contained finer blocks. During manual labeling, these finer blocks may have been distinguished by the naked eye, while the SAM may not have been able to segment them. Moreover, overlaps or shading between small blocks may also cause the SAM to generate a larger block mask during segmentation. Except for the five characteristic sizes mentioned above, the results for the rest of the characteristic sizes ( X 60 , X 70 , X 80 , X 90 , and X 100 ) were highly consistent with the manual segmentation results, as shown in Figure 10. Figure 11 displays the evaluation results for the obtained 10 characteristic sizes. The R2 values for all the characteristic sizes except X 10 and X 20 were higher than 0.95, and the RMSE and MAE values were less than 0.1, which indicates that the SAM could provide accurate measurements of rock fragment size.

4.2. Performance of the SAM on Poor-Quality Images

In this section, the segmentation performance of the SAM on the poor-quality images is discussed. The sharpness values of these images ranged from 6 to 14, and the total number of images was 30. Most of the poor-quality images were obtained under poor environmental conditions, such as poor brightness and contour texture, more dust or dirt, uneven block sizes, and so on. The collected images are shown in Figure 12.
(1)
Pixel-level segmentation results
Figure 13 displays the segmentation results of the SAM on some poor-quality rock block images. It can be seen that the performance of the SAM was affected to some extent. For example, in image 1 and image 2 of Figure 13, the background color of the test samples is darker, which resulted in the fragments in the red box region not being well recognized. In addition, the surfaces of the rock blocks in images 3 to 8 are covered with dirt or dust, making the outlines of the rock blocks in the images unclear. As a result, it was easy to miss or misidentify certain rock blocks, e.g., multiple small blocks may have been merged into a large block, as shown in the green boxed area in Figure 13. For image 9 in Figure 13, the blocks in the boxed area are also not segmented due to the poor lighting condition and contour tracks.
Although the segmentation performance of the SAM was somewhat affected in the poor-quality images, it only performed poorly on a few rock blocks in localized areas. And the values of the three evaluation indicators were 0.936, 0.935, and 0.948, respectively. Compared with the performance on high-quality data, there was only a decrease of 1.9%, 1.8%, and 1.2%, respectively, as shown in Table 3.
(2)
Block size distribution results
In addition, 10 characteristic size values for all poor-quality images were obtained using the SAM, as shown in Figure 14. The characteristic size results from X 10 to X 40 showed some differences compared to the manual segmentation results. However, the results for the other characteristic sizes were unaffected, indicating that the image quality had no significant impact on the final results for the primary characteristic sizes, such as X 50 and X 80 . Furthermore, the 10 characteristic sizes generated from the SAM and manual segmentation results were compared based on the three evaluation indexes, R2, RMSE, and MRE, as shown in Figure 15. The R2 value of the characteristic size larger than X 50 exceeded 0.95, and the RMSE and MRE were both less than 0.1. This demonstrated that the SAM method applied in this paper could obtain satisfactory performance even on poor-quality images.

5. Analysis and Discussion

In this section, the effectiveness and the generalizability of the SAM for rock fragmentation segmentation are discussed. First, the effectiveness of the SAM was evaluated by comparing it with commercial software (Split Desktop). Second, the generalizability of the SAM model was verified using two additional datasets collected from the previous literature.

5.1. Comparison with the Split Desktop

Split Desktop is a well-known commercial software for rock block size prediction and analysis. It is frequently utilized in real engineering applications or as an auxiliary tool for theoretical analysis in academic research. Figure 16 and Figure 17 show the segmentation results of the two types of images using the Split Desktop software and the SAM, respectively.
(1)
Comparison of pixel-level segmentation results
As shown in Figure 16, the SAM achieved comparable results to the Split Desktop software for high-quality images. Additionally, it is worth noting that the SAM method even outperformed the Split Desktop software for larger blocks or blocks with sharp contours. For example, in the areas marked by the red boxes in image 1 to image 3, the surface of these blocks is shaded; the Split Desktop software performed under segmentation, making the shadowed regions classified as the finer particles. For image 4 in Figure 16, the two blocks marked by the red boxes have sharp surface contours, and the Split Desktop software mistakenly segmented them into several small blocks, leading to over-segmentation and making the results smaller than the true values. For these cases, however, the SAM method performed extremely robustly, and all the marked blocks were accurately segmented.
Figure 17 exhibits the comparison results of the poor-quality samples. As can be seen from images 5 to 10, there were more defective segmentation objects marked in the red boxes, and many of them had low brightness, poor texture features, and dusty or dirty surfaces. The Split Desktop software failed to provide satisfactory segmentation results for such fragments, and it directly segmented them into several smaller blocks. Compared to the Split Desktop software, the SAM could give more accurate segmentation results and displayed stronger generalization capabilities.
(2)
Comparison of block size distribution results
Apart from that, the fragment size distribution results obtained by the Split Desktop and the SAM were compared, as shown in Figure 18 and Table 4. Figure 18 exhibits the accumulated percent passing results obtained by the Split Desktop on four high-quality test images. Table 4 shows the results for the 10 characteristic sizes. Obviously, the characteristic sizes X 10 , X 20 , X 30 , and X 40 obtained from the Split Desktop were smaller than the real values. This was mainly due to the under-segmentation of the Split Desktop, which made some blocks not fully segmented. In addition, the difference in the segmentation results between the Split Desktop and the true values tended to increase gradually in characteristic size ranges from X 50 to X 90 . The main reason for this discrepancy was the over-segmentation of the Split Desktop, i.e., some large blocks were split into smaller ones. Because of this, the number of blocks with characteristic sizes between X 50 and X 90 increased, leading to large fluctuations in the corresponding values. At the same time, the value of the characteristic size X 100 was naturally smaller than the true value due to the incorrect segmentation of some large blocks into small ones. The segmentation performance of the SAM method was quite stable on the four high-quality testing images, as shown in Figure 10 and Figure 11.
The block size distribution comparisons between the Split Desktop and the SAM methods on all the high-quality images and the poor-quality images are summarized in Figure 19. The mean absolute error (MAE) was used to evaluate the accuracy of the segmentation results, as defined in Equation (10):
M A E = 1 M i = 1 M y i y i
where y i and y i represent the target values and the segmentation results, respectively. M is the number of testing samples.
The bars marked with red asterisks indicate the results obtained from Split Desktop, while the bars without an asterisk represented the results from the SAM. As can be seen in Figure 19, the mean absolute errors of the 10 characteristic sizes obtained by Split Desktop were higher than those obtained by the SAM method for both high-quality and low-quality data. The SAM performed more stably; the average MAE value for all characteristic sizes was less than 1. Additionally, for Split Desktop, the mean absolute errors exhibited a “U” shape from characteristic size X 10 to X 100 in both high-quality and poor-quality images, achieving the best result at X 50 . According to Figure 19a,b, the mean absolute errors of all the characteristic sizes were very close to each other on both high-quality and poor-quality images, indicating that Split Desktop was also a well-developed and robust software.

5.2. Performance on Another Dataset

Furthermore, two other datasets presented in previous studies were collected to verify the generalizability of the SAM. In the study of Guo et al. [18], the authors developed an improved watershed algorithm to segment blasted rock images; two test images were used in their study to evaluate the performance of the modified watershed method. As shown in Figure 20, the two test images were also segmented by the SAM method. The developed watershed algorithm segmented almost all of the fragments, from large blocks to very small parts. However, there were some apparent errors in the second test image, as shown in the boxed region of Figure 20. Traditional image segmentation methods involve various image preprocessing steps (such as image binarization, denoising, and morphology process), and the adjusted threshold values only worked for the specified data. In contrast, the SAM applied in this paper achieved superior performance on the two test images.
The second validation test was conducted based on the data used in another study [21]. In that study, they created a deep learning-based approach to measure the size of rock fragments by combining different neural network structures. It is worth noting that the data preparation and image labeling tasks are the most time-consuming and labor-intensive tasks for training deep learning models. Nevertheless, this is an indispensable step. Consequently, a total of 14,628 images with 512 × 512 size were prepared as data in their study, and two images with sizes of 2304 × 2016 and 1408 × 2080 were used as the test data. The two testing images were applied in this paper, and the values of the 10 characteristic sizes were calculated, as shown in Figure 21. The dotted lines with square markers indicate the SAM segmentation results, the dotted lines with triangle markers are the segmentation results of the model developed by other scholars, and the dotted lines with circle markers represent the true values. In addition, Figure 22 shows the segmentation results of the test images obtained using the SAM and the previous study; the results demonstrate that the SAM could achieve comparable accuracy to the method proposed in the previous study. Moreover, the biggest advantage of the SAM is that it did not require training data for final segmentation and exhibited excellent zero-shot segmentation performance.

6. Conclusions

This paper applied a deep learning-based image segmentation fundamental method, namely, the Segment Anything Model (SAM), to achieve the automatic segmentation of rock fragments. The performance of the SAM was analyzed in detail regarding pixel-level segmentation accuracy and size distribution values. The main conclusions are as follows:
(1)
In terms of pixel-level segmentation, the SAM achieved high accuracy on all test data, with average values of 94.5%, 94.4%, and 95.4% for PA, IOU, and Dice, respectively. In addition, the R2 values of the measured characteristic sizes X50 and X80 reached 0.97 and 0.99, respectively. The results indicate that the SAM exhibits excellent zero-shot performance on rock fragments’ segmentation.
(2)
The SAM demonstrated strong robustness and outperformed the commercial software (Split Desktop), particularly on blocks with low brightness, poor texture features, and dusty or dirty surfaces. The average MAE values for the SAM and Split Desktop on the test data were 0.936 and 6.154, respectively. Furthermore, on the two other datasets, the SAM achieved comparable accuracy to the traditional image segmentation method and deep learning-based method developed by other scholars.
(3)
Overall, the SAM produced satisfactory results in segmenting rock fragments. Based on these accurate measurements, the rapid evaluation of blasting operations and precise optimization of blasting parameters can be achieved. Moreover, potential applications of the SAM in mining and geotechnical engineering could be explored, such as mineral composition analysis and measurement of rock quality designation (RQD) based on digital images of core holes.

7. Limitations

Although the SAM model demonstrates high precision in rock fragment segmentation, our study has certain limitations. Due to the constraints associated with data collection post-blasting, we assumed that the surface rock sizes provided a representative sample of the overall rock size distribution within the pile. This assumption may not always hold true, particularly in cases of significant variation in rock size distribution within the pile. Future research should explore more comprehensive sampling methods to enhance the reliability of the findings.

Author Contributions

Conceptualization, J.Z.; modeling, J.Z.; writing—original draft, J.Z.; supervision, D.L.; writing—review and editing, D.L.; field work, Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The present research was financially supported by the National Natural Science Foundation of China (52374153).

Data Availability Statement

Data will be made available upon request.

Conflicts of Interest

Yisong Yu is employees of CINF Engineering Co., Ltd. The paper reflects the views of the scientists and not the company.

References

  1. Shehu, S.A.; Yusuf, K.O.; Hashim, M. Comparative study of WipFrag image analysis and Kuz-Ram empirical model in granite aggregate quarry and their application for blast fragmentation rating. Geomech. Geoeng. 2022, 17, 197–205. [Google Scholar] [CrossRef]
  2. Inanloo Arabi Shad, H.; Sereshki, F.; Ataei, A.; Karamoozian, M. Investigation of the rock blast fragmentation based on the specific explosive energy and in-situ block size. Int. J. Min. Geo-Eng. 2018, 52, 1–6. [Google Scholar] [CrossRef]
  3. Kinyua, E.M.; Zhang, J.H.; Kasomo, R.M.; Mauti, D.; Mwangangi, J. A review of the influence of blast fragmentation on downstream processing of metal ores. Min. Eng. 2022, 186, 107743. [Google Scholar] [CrossRef]
  4. Amoako, R.; Jha, A.; Zhong, S. Rock fragmentation prediction using an artificial neural network and support vector regression hybrid approach. Mining 2022, 2, 233–247. [Google Scholar] [CrossRef]
  5. Cunningham, C.V.B. The KuzRam Model for Prediction of Fragmentation from Blasting. In Proceedings of the First International Symposium on Rock Fragmentation by Blasting, Lulea, Sweden, 22–26 August 1983; Luleå University of Technology: Lulea, Sweden, 1983; pp. 439–453. [Google Scholar]
  6. Kanchibotla, S.S.; Valery, W.; Morrell, S. Modelling fines in blast fragmentation and its impact on crushing and grinding. In Proceedings of the Explo ‘99—A Conference on Rock Breaking, Kalgoorlie, WA, USA, 7–11 November 1999; The Australasian Institute of Mining and Metallurgy: Kalgoorlie, Australia, 1999; pp. 137–144. [Google Scholar]
  7. Djordjevic, N. Two-component model of the blast fragmentation. In Proceedings of the 6th International Symposium for Rock Fragmentation by Blasting-Fragblast, Johannesburg, South Africa, 8–12 August 1999; pp. 213–219. [Google Scholar]
  8. Ouchterlony, F. The Swebrec© function: Linking fragmentation by blasting and crushing. Min. Technol. 2005, 114, 29–44. [Google Scholar] [CrossRef]
  9. Kou, S.; Rustan, A. Computerized Design and Result Prediction of Bench Blasting. In Proceedings of the Fourth International Symposium on Rock Fragmentation by Blasting, Vienna, Austria, 5–8 July 1993; pp. 263–271. [Google Scholar]
  10. Maerz, N.H.; Palangio, T.C.; Franklin, J.A. WipFrag image based granulometry system. In Proceedings of the Fragblast-5 Workshop on Measurement of Blast Fragmentation, Montreal, QC, Canada, 23–24 August 1996; pp. 91–98. [Google Scholar]
  11. Kemeny, J.M. Practical technique for determining the size distribution of blasted benches waste dump and heap leach sites. Min. Eng. 1994, 46, 1281–1284. [Google Scholar]
  12. Schleifer, J.; Tessier, B. FragScan, a tool to measure fragmentation of blasted rock. In Proceedings of the Fragblast-5 Workshop on Measurement of Blast Fragmentation, Montreal, QC, Canada, 23–24 August 1996; pp. 73–78. [Google Scholar]
  13. Raina, A.K.; Choudhary, P.B.; Ramulu, M.; Chakraborty, A.K.; Dudhankar, A.S.; Udpikar, V.; Ghatpande, N.; Misra, D.D. Fragalyst-An indigenous digital image analysis system for fragment size measurement in mines. J. Geol. Soc. India 2002, 59, 561–569. [Google Scholar]
  14. Chung, S.H.; Noy, M.J. Experience in fragmentation control. In Proceedings of the Fragblast-5 Fragblast Workshop on Measurement of Blast Fragmentation, Montreal, QC, Canada, 23–24 August 1996; Balkema: Rotterdam, The Netherlands; pp. 247–252. [Google Scholar]
  15. Kleine, T.H.; Cameron, A.R. Blast fragmentation measurement using Goldsize. In Proceedings of the Fragblast-5 Workshop on Measurement of Blast Fragmentation, Montreal, QC, Canada, 23–24 August 1996; Balkema: Rotterdam, The Netherlands; pp. 83–89. [Google Scholar]
  16. Amankwah, A.; Aldrich, C. Rock image segmentation using watershed with shape markers. In Proceedings of the 2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 13–15 October 2010; pp. 1–7. [Google Scholar] [CrossRef]
  17. Sereshki, F.; Hoseini, S.M.; Ataei, M. Blast fragmentation analysis using image processing. Int. J. Min. Geo-Eng. 2016, 50, 211–218. [Google Scholar] [CrossRef]
  18. Guo, Q.P.; Wang, Y.C.; Yang, S.J.; Xiang, Z.B. A method of blasted rock image segmentation based on improved watershed algorithm. Sci. Rep. 2022, 12, 7143. [Google Scholar] [CrossRef]
  19. Yang, Z.; Ding, H.J.; Guo, L.; Lian, M.J. Superpixel image segmentation-based particle size distribution analysis of fragmented rock. IEEE Access 2021, 9, 59048–59058. [Google Scholar] [CrossRef]
  20. Wang, W.; Li, Q.; Xiao, C.Y.; Zhang, D.Z.; Miao, L.; Wang, L. An improved boundary-aware U-Net for ore image semantic segmentation. Sensors 2021, 21, 2615. [Google Scholar] [CrossRef] [PubMed]
  21. Guo, Q.P.; Yang, S.J.; Wang, Y.C.; Zhang, H.B. Size measurement of blasted rock fragments based on FRRSnet+. Measurement 2023, 218, 113207. [Google Scholar] [CrossRef]
  22. Qassim, H.; Verma, A.; Feinzimer, D. Compressed residual-VGG16 CNN model for big data places image recognition. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–10 January 2018; pp. 169–175. [Google Scholar]
  23. Sandler, M.; Howard, A.; Zhu, M.L.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  24. Ma, N.N.; Zhang, X.Y.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 122–138. [Google Scholar]
  25. Li, D.Y.; Zhao, J.J.; Ma, J.Y. Experimental Studies on Rock Thin-Section Image Classification by Deep Learning-Based Approaches. Mathematics 2022, 10, 2317. [Google Scholar] [CrossRef]
  26. Ren, S.Q.; He, K.M.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems 28, Proceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; The Neural Information Processing Systems Foundation (NIPS): La Jolla, CA, USA, 2015; pp. 91–99. [Google Scholar]
  27. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot MultiBox detector. In Computer Vision—ECCV 2016, Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  28. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  29. Li, D.Y.; Zhao, J.J.; Liu, Z.D. A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks. Sensors 2022, 22, 1574. [Google Scholar] [CrossRef] [PubMed]
  30. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  31. He, K.M.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  32. Huang, H.M.; Lin, L.F.; Tong, R.F.; Hu, H.J.; Zhang, Q.W.; Iwamoto, Y.; Han, X.; Chen, Y.W.; Wu, J. UNet 3+: A full-scale connected UNet for medical image segmentation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing 2020, Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  33. Li, Y.; Duan, Y.Y. A method of ore image segmentation based on deep learning. In Proceedings of the International Conference on Intelligent Computing, Wuhan, China, 15–18 August 2018; pp. 508–519. [Google Scholar]
  34. Qiao, W.D.; Zhao, Y.F.; Xu, Y.; Lei, Y.M.; Wang, Y.J.; Yu, S.; Li, H. Deep learning-based pixel-level rock fragment recognition during tunnel excavation using instance segmentation model. Tunn. Undergr. Space Technol. 2021, 115, 104072. [Google Scholar] [CrossRef]
  35. Zhou, X.X.; Gong, Q.M.; Liu, Y.Q.; Yin, L.J. Automatic segmentation of TBM muck images via a deep-learning approach to estimate the size and shape of rock chips. Autom. Constr. 2021, 126, 103685. [Google Scholar] [CrossRef]
  36. Li, X.; Li, S.Y.; Dong, L.; Su, S.X.; Hu, X.J.; Lu, Z.L. An image segmentation method of pulverized coal for particle size analysis. Int. J. Min. Sci. Technol. 2023, 33, 1181–1192. [Google Scholar] [CrossRef]
  37. Bamford, T.; Esmaeili, K.; Schoellig, A.P. A deep learning approach for rock fragmentation analysis. Int. J. Rock. Mech. Min. Sci. 2021, 145, 104839. [Google Scholar] [CrossRef]
  38. Li, H.; Asbjörnsson, G.; Lindqvist, M. Image process of rock size distribution using dexined-based neural network. Minerals 2021, 11, 736. [Google Scholar] [CrossRef]
  39. Poma, X.S.; Riba, E.; Sappa, A. Dense extreme inception network: Towards a robust cnn model for edge detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass, CO, USA, 1–5 March 2020; pp. 1923–1932. [Google Scholar]
  40. Duan, J.X.; Liu, X.Y.; Wu, X.; Mao, C.G. Detection and segmentation of iron ore green pellets in images using lightweight U-net deep learning network. Neural Comput. Appl. 2020, 32, 5775–5790. [Google Scholar] [CrossRef]
  41. Mustafa, N.; Zhao, J.P.; Liu, Z.Y.; Zhang, Z.H.; Yu, W.X. Iron ORE region segmentation using high-resolution remote sensing images based on Res-U-Net. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2563–2566. [Google Scholar]
  42. Yang, Z.; Wu, H.; Ding, H.J.; Liang, J.M.; Guo, L. Enhanced U-Net model for rock pile segmentation and particle size analysis. Min. Eng. 2023, 203, 108352. [Google Scholar] [CrossRef]
  43. Jin, C.Y.; Liang, J.Y.; Fan, C.H.; Chen, L.J.; Wang, Q.; Lu, Y.; Wang, K. Study on segmentation of blasting fragment images from open-pit mine based on U-CARFnet. PLoS ONE 2023, 18, e0291115. [Google Scholar] [CrossRef] [PubMed]
  44. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.Z.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment anything. arXiv 2023, arXiv:2304.02643. [Google Scholar] [CrossRef]
  45. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar] [CrossRef]
  46. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. In Proceedings of the 34th Conference on Neural Information Processing Systems, Vancouver, BC, Canada/Online Event, 6–12 December 2020; Volume 33, pp. 1877–1901. [Google Scholar]
  47. OpenAI; Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; et al. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
  48. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  49. He, K.M.; Chen, X.L.; Xie, S.N.; Li, Y.H.; Dollar, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar]
  50. Ahmadi, M.; Lonbar, A.G.; Sharifi, A.; Beris, A.T.; Nouri, M.; Javidi, A.S. Application of segment anything model for civil infrastructure defect assessment. arXiv 2023, arXiv:2304.12600. [Google Scholar] [CrossRef]
  51. Yang, J.Y.; Gao, M.Q.; Li, Z.; Gao, S.; Wang, F.J.; Zheng, F. Track anything: Segment anything meets videos. arXiv 2023, arXiv:2304.11968. [Google Scholar] [CrossRef]
  52. Yu, T.; Feng, R.S.; Feng, R.Y.; Liu, J.M.; Jin, X.; Zeng, W.J.; Chen, Z.B. Inpaint anything: Segment anything meets image inpainting. arXiv 2023, arXiv:2304.06790. [Google Scholar] [CrossRef]
  53. Xie, D.F.; Wang, R.C.; Ma, J.; Chen, C.; Lu, H.N.; Yang, D.; Shi, F.B.; Lin, X.D. Edit everything: A text-guided generative system for images editing. arXiv 2023, arXiv:2304.14006. [Google Scholar] [CrossRef]
  54. Liu, S.H.; Ye, J.W.; Wang, X.C. Any-to-any style transfer: Making picasso and da vinci collaborate. arXiv 2023, arXiv:2304.09728. [Google Scholar] [CrossRef]
  55. Ma, J.; He, Y.T.; Li, F.F.; Han, L.; You, C.Y.; Wang, B. Segment anything in medical images. arXiv 2023, arXiv:2304.12306. [Google Scholar] [CrossRef] [PubMed]
  56. Cheng, D.J.; Qin, Z.Y.; Jiang, Z.K.; Zhang, S.T.; Lao, Q.C.; Li, K. Sam on medical images: A comprehensive study on three prompt modes. arXiv 2023, arXiv:2305.00035. [Google Scholar] [CrossRef]
  57. Mazurowski, M.A.; Dong, H.Y.; Gu, H.X.; Yang, J.C.; Konz, N.; Zhang, Y.X. Segment anything model for medical image analysis: An experimental study. Med. Image Anal. 2023, 89, 102918. [Google Scholar] [CrossRef] [PubMed]
  58. Chung, S.H.; Katsabanis, P.D. Fragmentation prediction using improved engineering formulae. Fragblast 2000, 4, 198–207. [Google Scholar] [CrossRef]
  59. Sudhakar, J.; Adhikari, G.R.; Gupta, R.N. Comparison of fragmentation measurements by photographic and image analysis techniques. Rock. Mech. Rock. Eng. 2006, 39, 159–168. [Google Scholar] [CrossRef]
Figure 1. Zero-shot segmentation performance of the SAM on the open dataset [44].
Figure 1. Zero-shot segmentation performance of the SAM on the open dataset [44].
Minerals 14 00654 g001
Figure 2. Flowchart of the SAM [44].
Figure 2. Flowchart of the SAM [44].
Minerals 14 00654 g002
Figure 3. Examples of the collection dataset. The red pen and the flashlight are size calibrators.
Figure 3. Examples of the collection dataset. The red pen and the flashlight are size calibrators.
Minerals 14 00654 g003
Figure 4. Calculation results of the sharpness and blurriness for all images.
Figure 4. Calculation results of the sharpness and blurriness for all images.
Minerals 14 00654 g004
Figure 5. The loss results of image clustering based on sharpness and blurriness, respectively.
Figure 5. The loss results of image clustering based on sharpness and blurriness, respectively.
Minerals 14 00654 g005
Figure 6. The classification results for the sharpness and blurriness metrics.
Figure 6. The classification results for the sharpness and blurriness metrics.
Minerals 14 00654 g006
Figure 7. SAM segmentation performance on the four test images: (a) original images, (b) manual segmentation results, and (c) SAM segmentation results.
Figure 7. SAM segmentation performance on the four test images: (a) original images, (b) manual segmentation results, and (c) SAM segmentation results.
Minerals 14 00654 g007
Figure 8. Block size distributions of the four testing samples: (ad) are the test image 1 to image 4, respectively.
Figure 8. Block size distributions of the four testing samples: (ad) are the test image 1 to image 4, respectively.
Minerals 14 00654 g008
Figure 9. Comparison of the accumulated passing percentage between the SAM and manual segmentation results on four testing samples: (ad) are image 1–image 4, respectively.
Figure 9. Comparison of the accumulated passing percentage between the SAM and manual segmentation results on four testing samples: (ad) are image 1–image 4, respectively.
Minerals 14 00654 g009
Figure 10. Ten characteristic sizes of all high-quality images obtained by the SAM.
Figure 10. Ten characteristic sizes of all high-quality images obtained by the SAM.
Minerals 14 00654 g010
Figure 11. The results of R2, RMSE, and MAE for different characteristic sizes on high-quality images.
Figure 11. The results of R2, RMSE, and MAE for different characteristic sizes on high-quality images.
Minerals 14 00654 g011
Figure 12. Samples of poor-quality dataset.
Figure 12. Samples of poor-quality dataset.
Minerals 14 00654 g012
Figure 13. SAM segmentation performance on poor-quality images: (a,d) original images, (b,e) labeled results, and (c,f) SAM segmentation results.
Figure 13. SAM segmentation performance on poor-quality images: (a,d) original images, (b,e) labeled results, and (c,f) SAM segmentation results.
Minerals 14 00654 g013aMinerals 14 00654 g013b
Figure 14. Ten characteristic sizes of all poor-quality images obtained by the SAM.
Figure 14. Ten characteristic sizes of all poor-quality images obtained by the SAM.
Minerals 14 00654 g014
Figure 15. The results of R2, RMSE, and MAE for different characteristic sizes on poor-quality images.
Figure 15. The results of R2, RMSE, and MAE for different characteristic sizes on poor-quality images.
Minerals 14 00654 g015
Figure 16. Segmentation results for high-quality images: (a) original images, (b) Split Desktop segmentation results, and (c) SAM segmentation results.
Figure 16. Segmentation results for high-quality images: (a) original images, (b) Split Desktop segmentation results, and (c) SAM segmentation results.
Minerals 14 00654 g016
Figure 17. Segmentation results for poor-quality images: (a) original images, (b) Split Desktop segmentation results, and (c) SAM segmentation results.
Figure 17. Segmentation results for poor-quality images: (a) original images, (b) Split Desktop segmentation results, and (c) SAM segmentation results.
Minerals 14 00654 g017
Figure 18. Size distribution results of the four testing images obtained by Split Desktop.
Figure 18. Size distribution results of the four testing images obtained by Split Desktop.
Minerals 14 00654 g018aMinerals 14 00654 g018b
Figure 19. Mean absolute errors obtained from the Split Desktop and the SAM methods: (a) the results for the high-quality images and (b) the results for the poor-quality images.
Figure 19. Mean absolute errors obtained from the Split Desktop and the SAM methods: (a) the results for the high-quality images and (b) the results for the poor-quality images.
Minerals 14 00654 g019
Figure 20. Comparison of segmentation results between the SAM and the other method: (a) original images, (b) manual labeling results, (c) segmentation results from previous study [18], (d) SAM segmentation results.
Figure 20. Comparison of segmentation results between the SAM and the other method: (a) original images, (b) manual labeling results, (c) segmentation results from previous study [18], (d) SAM segmentation results.
Minerals 14 00654 g020
Figure 21. Comparison of 10 characteristic size values obtained by the SAM and the previous study [21] on two test images: (a) image 1, (b) image 2.
Figure 21. Comparison of 10 characteristic size values obtained by the SAM and the previous study [21] on two test images: (a) image 1, (b) image 2.
Minerals 14 00654 g021
Figure 22. The comparison of segmentation results between SAM and the previous study: (a) original images, (b) segmentation results from previous study [21], (c) SAM segmentation results.
Figure 22. The comparison of segmentation results between SAM and the previous study: (a) original images, (b) segmentation results from previous study [21], (c) SAM segmentation results.
Minerals 14 00654 g022
Table 1. Quantitative evaluation of SAM pixel-level segmentation performance on high-quality images.
Table 1. Quantitative evaluation of SAM pixel-level segmentation performance on high-quality images.
Evaluation Metrics
SAMPAmIOUDice Coefficient
0.9550.9530.960
Table 2. Comparison results of 10 characteristic sizes obtained by the SAM and manual segmentation results of four testing images.
Table 2. Comparison results of 10 characteristic sizes obtained by the SAM and manual segmentation results of four testing images.
Characteristic Size Parameters (mm)
X 10 X 20 X 30 X 40 X 50 X 60 X 70 X 80 X 90 X 100
Image 1True value8.3013.2017.6522.0526.6431.6737.5044.8956.0472.74
SAM7.6812.4616.8721.2625.8931.0036.9644.5756.1567.5
Image 2True value13.3519.3324.3629.0933.8138.8044.4051.2361.13111.29
SAM13.4119.2624.1428.7133.2638.0543.4049.9259.33108.52
Image 3True value9.0813.4717.2420.8224.4428.3132.6838.0745.9669.23
SAM8.9813.0516.5019.7523.0026.4430.3035.0241.8868.08
Image 4True value9.3813.7317.4220.9124.4128.1332.3237.4544.9293.16
SAM8.4512.5516.0519.3922.7726.3730.4435.4642.8191.31
where X 10 , X 20 , X 30 , X 40 , X 50 , X 60 , X 70 , X 80 , X 90 , and X 100 represent the size of the fragments that pass through the sieve at 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, and 100%, respectively.
Table 3. Comparison of SAM pixel-level segmentation performance on high- and poor-quality images.
Table 3. Comparison of SAM pixel-level segmentation performance on high- and poor-quality images.
Performance on High-Quality ImagesPerformance on Poor-Quality ImagesDifference
PA0.9550.936 1.9%
mIOU0.9530.935 1.8%
Dice Coefficient0.9600.948 1.2%
Table 4. Comparison results of 10 characteristic sizes obtained by the SAM, Split Desktop, and manual segmentation results of four testing images.
Table 4. Comparison results of 10 characteristic sizes obtained by the SAM, Split Desktop, and manual segmentation results of four testing images.
Characteristic Size Parameters (mm)
X 10 X 20 X 30 X 40 X 50 X 60 X 70 X 80 X 90 X 100
Image 1True value8.3013.2017.6522.0526.6431.6737.5044.8956.0472.74
SAM7.6812.4616.8721.2625.8931.0036.9644.5756.1567.5
Split Desktop5.2010.2115.0419.8224.8129.5034.0039.1446.2170.12
Image 2True value13.3519.3324.3629.0933.8138.8044.4051.2361.13111.29
SAM13.4119.2624.1428.7133.2638.0543.4049.9259.33108.52
Split Desktop5.2011.0817.1223.3530.0036.2342.6550.0059.7489.56
Image 3True value9.0813.4717.2420.8224.4428.3132.6838.0745.9669.23
SAM8.9813.0516.5019.7523.0026.4430.3035.0241.8868.08
Split Desktop3.828.0512.5617.1322.1826.8331.2037.8445.9275.76
Image 4True value 9.3813.7317.4220.9124.4128.1332.3237.4544.9293.16
SAM8.4512.5516.0519.3922.7726.3730.4435.4642.8191.31
Split Desktop4.328.7213.1517.6822.2126.4230.8335.8442.7180.04
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, J.; Li, D.; Yu, Y. Identification of Rock Fragments after Blasting by Using Deep Learning-Based Segment Anything Model. Minerals 2024, 14, 654. https://doi.org/10.3390/min14070654

AMA Style

Zhao J, Li D, Yu Y. Identification of Rock Fragments after Blasting by Using Deep Learning-Based Segment Anything Model. Minerals. 2024; 14(7):654. https://doi.org/10.3390/min14070654

Chicago/Turabian Style

Zhao, Junjie, Diyuan Li, and Yisong Yu. 2024. "Identification of Rock Fragments after Blasting by Using Deep Learning-Based Segment Anything Model" Minerals 14, no. 7: 654. https://doi.org/10.3390/min14070654

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop