Next Article in Journal
Monocular Visual Pig Weight Estimation Method Based on the EfficientVit-C Model
Previous Article in Journal
Efficiency of Backwashing in Removing Solids from Sand Media Filters for Drip Irrigation Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrated Scale-Adaptive Adjustment Factor-Enhanced BlendMask Method for Pineapple Processing System

1
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
2
Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
3
Institute of Physics, Chinese Academy of Sciences, Beijing 100080, China
*
Author to whom correspondence should be addressed.
Agriculture 2024, 14(9), 1569; https://doi.org/10.3390/agriculture14091569
Submission received: 7 August 2024 / Revised: 3 September 2024 / Accepted: 7 September 2024 / Published: 10 September 2024
(This article belongs to the Section Agricultural Technology)

Abstract

:
This study addresses the challenge of efficiently peeling pineapples, which have a distinct elliptical form, thick skin, and small eyes that are difficult to detect with conventional automated methods. This results in significant flesh waste. To improve the process, we developed an integrated system combining an enhanced BlendMask method, termed SAAF-BlendMask, and a Pose Correction Planning (PCP) method. SAAF-BlendMask improves the detection of small pineapple eyes, while PCP ensures accurate posture adjustment for precise path planning. The system uses 3D vision and deep learning technologies, achieving an average precision (AP) of 73.04% and a small object precision (APs) of 62.54% in eye detection, with a path planning success rate reaching 99%. The fully automated electromechanical system was tested on 110 real pineapples, demonstrating a reduction in flesh waste by 11.7% compared to traditional methods. This study highlights the potential of advanced machine vision and robotics in enhancing the efficiency and precision of food processing.

1. Introduction

In recent years, the global production of pineapples has shown a continuous upward trend. It was estimated that in 2020, approximately 40% of the global pineapple production (around 10 billion pineapples) was utilized for processing, yielding canned pineapples, pineapple juice, and other derivative products [1]. However, during processing, especially when dealing with the pineapple’s distinctive elliptical shape, thick outer skin, and deeply recessed pineapple eyes, the inability to accurately locate the pineapple eyes results in imprecise automated processing. Consequently, pineapples processed by machinery resembling the Ginaca machine [2,3] suffer from pronounced pulp wastage. Cylindrical cutting leads to significant wastage, as illustrated in Figure 1a. Manual handling using specialized tools has emerged as a common approach to address non-standard processing, offering flexibility and processing precision, as depicted in Figure 1b, illustrating the manual processing of pineapples. However, it requires a considerable amount of manual labor. The key aspects of automated pineapple processing include the precise identification of pineapple eyes and their rapid and accurate removal.
Artificial intelligence, especially deep learning, has revolutionized the field of fruit recognition and classification for the past few years. Technologies like Convolutional Neural Networks (CNNs) have been widely used for visual fruit recognition [4,5,6,7,8]. For instance, Zhang et al. [9] developed a five-layer CNN model for fruit classification. This model was applied to both the UEC-FOOD100 dataset [10] and an in-house fruit dataset. Liu et al. [11] utilized a YOLO-v3 [12] single-stage detector, enhancing it to accurately predict circular regions for tomato localization in intricate scenes. Yu et al. [13] introduced a Mask R-CNN [14] algorithm for recognizing 100 wild strawberry images. The results indicated an average recognition accuracy of 95.78% and a recall rate of 95.41%. In 2021, Wang et al. [15] proposed a modified VGG16 [16] network to count apple flowers in images. The total flower count represented the image class, and the network was trained to estimate the visible flower count without localization. Chen et al. [17] proposed another deep learning approach to perform segmentation, rather than detection, using a DCNN.
However, there is a significant scale difference between the pineapple contour and the pineapple eyes, resulting in low detection performance for small target pineapple eyes. Moreover, during the peeling process, excessive removal of the peel results in particularly small retained pineapple eye sizes, exacerbating the difficulty of small target recognition. To address the challenge of small object detection, researchers have proposed various methods, which can be categorized as follows: data augmentation methods to address the scarcity and non-uniform distribution of small object data. Kisanta et al. [18] improved the recognition accuracy and segmentation accuracy of small objects by 7.1% and 9.7%, respectively, through oversampling and augmentation. The Mosaic data augmentation method proposed in YOLOv4 [19] randomly splices four images together by flipping, scaling, and cropping, reducing the computational load on the graphics processing unit (GPU), but potentially decreasing the model’s generalization ability due to smaller object sizes. For super-resolution methods addressing the weak visual features of small objects, Haris et al. [20] proposed an end-to-end super-resolution method that uses the Faster R-CNN network to process low-resolution regions, improving small object detection performance. Bai et al. [21] proposed the multitask generative adversarial network (MTGAN), which, based on super-resolution methods, can up-sample blurry images to generate fine images corresponding to the desired scale for more accurate detection. For multi-scale fusion methods addressing the weak representation capability of single feature layers for small objects, Lin et al. [22] proposed feature pyramid networks (FPN), which integrate low-resolution strong semantic features with high-resolution weak semantic features to enhance the information of feature maps, fully exploiting the semantic information of each different scale feature map, thereby providing certain assistance for small object detection. However, during the fusion process, there may be issues with semantic gaps and noise, which to some extent hinder the further improvement of small object detection methods in terms of performance.
Pineapple processing has also gone through different stages. Traditional pineapple processing methods involve manual operations that require skilled labor to use various tools for tasks such as peeling, coring, and removing eyes. While manual peeling and slicing tools, as described in references [23,24,25] exist, they are challenging to operate and have low efficiency. To address labor challenges, automated processing methods, including mechanical peeling and coring machines, have been introduced. These methods have evolved over time and encompass various technologies. For instance, Hawaiian Pineapple Company (HPC) hired Henry Ginaca in 1911 developing a machine (Ginaca machine) that produced pineapple cylinders at a much higher rate [2]. Ginaca machines boast high efficiency but employ a relatively coarse processing method, resulting in significant fruit loss. Ref. [26] designed a semi-automatic peeling principle machine, with a particular focus on the stability of pineapple positioning. However, this machine still employs cylindrical blades, leading to substantial fruit loss during peeling. Ref. [27] designed a semi-automatic peeling principle machine that used a new process, namely rotary peeling, and conducted theoretical analysis along with experimental verification. Although these methods have relatively higher efficiency in peeling, they still suffer from issues like lower overall efficiency, the inability to remove eyes, and the need for manual eye removal, or imprecise processing that results in significant loss of pineapple pulp.
To date, there have been no complete studies on the automatic processing system of pineapple with integrated machine vision recognition. In this study, we propose an integrated approach for pineapple eye removal, incorporating a scale-adaptive adjustment factor and pose correction planning method. Additionally, we design and develop an innovative 3D vision-based intelligent motor-integrated pineapple processing system. This system achieves accurate identification and efficient automation of pineapple fruit processing, providing valuable insights into the automation of agricultural product processing.
The key contributions of this research are as follows:
(1)
We introduce a Scale-Adaptive Adjustment Factor (SAAF) to enhance the effectiveness of the BlendMask target segmentation algorithm, leading to significant improvements in detecting small targets. Multiple experiments have demonstrated that our algorithm exhibits robust performance and strong adaptability across different datasets.
(2)
We propose a Pose Correction Planning Method for the precise cutting path of pineapple eyes, ensuring a high success rate in path planning. This method achieves a path planning success rate of 99%.
(3)
We develop a fully automated, 3D vision-based intelligent motor-integrated pineapple processing system. The proposed algorithms are rigorously validated, demonstrating an average reduction in fruit wastage of 11.7% compared to traditional automated processing methods, such as the Ginaca machine [2].

2. Materials and Methods

As shown in Figure 2, the 3D vision-based mechatronics integrated pineapple processing system consists of the pineapple clamping platform mechanism, a peeling mechanism (which we constructed by referencing these work [25,28]), a 3D vision system, and an eye removal mechanism. The specific steps are as follows.
(1)
Data collecting—employing RGB-D camera to uniformly capture six images of each peeled pineapple, comprising both RGB imagery and depth information;
(2)
Image processing—utilizing an improved BlendMask algorithm (SAAF-BlendMask) for the identification of pineapples and pineapple eyes, followed by the reconstruction of the point cloud distribution of pineapple eyes;
(3)
Motion planning—planning of cutting paths (PCP method) and trajectories based on the distribution of pineapple eyes and cutting the pineapple eyes following the planned trajectories.
In particular, precise removal of pineapple eyes is critical and is achieved by emulating the manual removal process, particularly in steps (2) and (3). Accurate identification of pineapple eyes and the planning of removing trajectories are of paramount importance. The premise for trajectory planning is the alignment and maintenance of the pineapple’s central axis with the rotational axis, which is achieved in step (3). The following elaborate exposition will be presented.

2.1. Data Collecting

The pineapple needs to be peeled prior to image acquisition, as depicted in Figure 2. We have designed a peeling and clamping mechanism for this purpose. During the peeling process, the knife follows a spiral pattern on the pineapple’s surface, effectively removing the peel. This peeling mechanism maintains a relatively consistent peel thickness, typically ranging between 6 and 9 mm. This minimizes fruit wastage while retaining the pineapple eyes.
We opted for an RGB-D camera, specifically the Intel RealSense D435, to establish a depth network vision system. An RGB-D camera captures RGB images with depth values, or Z-depth, which are transformed into a 3D point cloud in the camera coordinate system O c using intrinsic parameters, and further converted to a global coordinate system O w using extrinsic parameters, as shown in Figure 3b.
Figure 3a provides a visual representation of the camera’s field of view. It is noteworthy that points at greater distances from O c exhibit higher errors (larger physical sizes corresponding to the same pixel), while points too close may become out of focus. Establishing an acceptable range is crucial to ensuring the uniform and accurate capture of points by the camera, as depicted in Figure 3a. Therefore, during camera installation, we aim to position the camera as close to the pineapple as possible within the effective detection range. This is to ensure that the pineapple eyes occupy a greater number of pixels.
Pineapple Image Acquisition. To enhance the effectiveness and robustness of our algorithm, we collected pineapple images from real-world scenarios.
Prior to image acquisition, the visual system was installed and calibrated with the processing equipment according to the technical specifications of the D435 camera (Intel Corporation, Santa Clara, CA, USA).
In order to reconstruct the complete point cloud distribution of pineapple eyes more accurately, multiple images of each pineapple were captured from different angles. Each image had a resolution of 640 × 480 pixels, and images were captured at 60-degree intervals. After every capture, the pineapple was rotated by 60 degrees. After six captures, the pineapple returned to its initial angle, ensuring consistency between the reconstructed pineapple pose and its actual pose.
Then, a significant number of images were collected and processed using data augmentation techniques, such as brightness adjustments, to account for varying lighting conditions in the processing environment. The collected images were further processed using the Labelme annotation software (version is 4.5.9, created by Kentaro Wada, and sourced from Berkeley, CA, USA).

2.2. SAAF-BlendMask Algorithm

The BlendMask algorithm belongs to one-stage dense instance segmentation methods. It is based on the anchor-free detection model FCOS [29] and combines instance-level information (represented by the orange dashed box in Figure 4) and finer-grained semantic information (represented by the green dashed box in Figure 4) to enhance mask prediction. It also introduces a blending module (represented by the yellow-green block in Figure 4), combining top-down and bottom-up instance segmentation methods. This effectively predicts position-sensitive instance features for dense pixel points using only one convolutional layer, preserving advantages in speed. In some cases, it outperforms Mask R-CNN in both speed and performance, meeting the real-time processing requirements of the system [30].
During reconstruction, each pineapple is reconstructed from the point cloud extracted from six images, with overlapping point clouds between adjacent images. Therefore, precision in point cloud extraction is particularly crucial. In the task of pineapple eye recognition and segmentation, segmentation networks can more accurately extract point cloud information relevant to pineapple eyes compared to traditional object detection methods (which typically use bounding boxes to label targets). Hence, we introduce the real-time instance segmentation algorithm BlendMask as the baseline.
To enhance the recognition performance of small-scale pineapple eyes and mitigate the influence of pineapple contour variations and eye size discrepancies, we devised a scale-adaptive balancing factor to refine the BlendMask algorithm [30]. The overall architecture of the vision system algorithm, as depicted in Figure 4, comprises two main components: instance segmentation and 3D point cloud reconstruction.
This algorithm employs a pyramid structure network (FPN), efficiently extracting feature maps at different resolutions, addressing the issue of target overlap. To address the challenge posed by the imbalance in size between pineapples and pineapple eyes in this application scenario, we proposed the Scale-Adaptive Adjustment Factor (SAAF) to enhance the BlendMask loss function, thereby improving the recognition accuracy of small target pineapple eyes.
In the FCOS architecture’s head module, pixel-wise target detection is performed using different hierarchical feature maps to predict targets, regressing 4D vectors T = ( l , b , r , t ) for the target bounding boxes to improve accuracy, as shown in Figure 5. To enhance accuracy further, the concept of centerness [29] is introduced to eliminate low-quality predicted boxes that are far from the center.
c e n t e r n e s s = m i n ( l , r ) m a x ( l , r ) m i n ( t , b ) m a x ( t , b ) .
To better address the issue of large-scale differences between pineapple eyes and pineapples leading to overlap, we conducted a detailed statistical analysis. The results indicated that the average resolution of pineapple eyes is about 20 × 20 pixels, while pineapples have an average resolution of about 200 × 250 pixels. Therefore, we adapted FCOS’s feature maps with five different sizes, denoted as P 3 , P 4 , P 5 , P 6 and P 7 , each with corresponding strides of 8, 16, 32, 64, and 128. We limited the regression distances differently for each feature map to achieve this purpose. Specifically, each feature map P i was assigned lower and upper bounds m i 1 and mi as constraints for regression distances. For each coordinate point ( x , y ) on feature map P i , if m a x ( l , t , r , b ) > m i or m i n ( l , t , r , b ) < m i 1 , it was treated as a negative sample on that feature layer, and regression was no longer performed. m 2 , m 3 , m 4 , m 5 , m 6 are set to 0, 64, 128, 256, and +, respectively. This implies that feature map P 3 handles target regression values within the [0, 64] range, while feature map P 7 is responsible for targets with regression values greater than 256.
In the previous analysis, although FCOS employs hierarchical feature maps to predict targets of different scales, the practical situation is that, on average, the number of anchor points in pineapple contour significantly exceed the number of anchor points on pineapple eyes. This imbalance leads to differences in centerness weight between anchor points on large pineapple targets and small pineapple eyes. Specifically, the centerness weight of small pineapple eyes tends to be lower, thus, they are often treated as low-quality regression boxes and suppressed. However, in the processing pipeline of this study, the peeling operation results in significant differences in the size of retained pineapple eyes, including very small pineapple eyes being quite common. Since the system’s goal is to accurately remove each residual pineapple eye, identifying these commonly occurring small pineapple eyes becomes crucial. To address the issue of weight imbalance caused by differences in target scales, we introduced the SAAF module to enhance the centerness. This enhancement aims to ensure that small pineapple eyes are equally well-recognized and processed, thereby improving the overall performance of the system.
The inputs of the SAAF module consist of ground truth and all positive samples, while the output is the adjustment factor. Initially, we compute the average area μ s g t of the ground truth,
μ s g t = i N g t s g t i / N g t ,
where s g t i represents the area of the i t h ground truth, N g t represents the number of ground truths. corresponding to the area s s a ( x , y ) of each positive sample (at the ( x , y ) point) from the ground truth.
s s a ( x , y ) = ( l + r ) ( b + t ) .
We determine the centerness of each positive sample and calculate a factor ω s ( x , y ) for each sample’s centerness.
ω s ( x , y ) = μ s g t s s a ( x , y ) ,
μ ω s ( x , y ) = x , y ω s ( x , y ) / N s ,
this is achieved by comparing against a threshold μ ω s ( x , y ) , where N s represents the number of positive samples for each ground truth.
k x , y = ω s ( x , y ) m , ( ω s ( x , y ) > μ ω s ( x , y ) ) 1 n , ( ω s ( x , y ) < μ ω s ( x , y ) ) ,
where m and n represent hyperparameters, with a default value of 1. Therefore, each positive sample’s weight is represented as s x , y = k x , y c e n t e r n e s s , which means it is adaptively adjusted based on the ground truth scale. After this adjustment, the loss function still consists of three parts: classification loss, regression loss, and centerness loss, combined.
L ( { p x , y } , { t x , y } , { s x , y } ) = 1 N s x , y L c l s ( p x , y , c x , y ) + 1 N s x , y 1 { s x , y > 0 , c x , y > 0 } L r e g ( t x , y , t x , y ) + 1 N s x , y 1 { c x , y > 0 } L c t r ( s x , y , s x , y ) ,
where p x , y represents the scores for each class predicted at the ( x , y ) point in the feature map. t x , y represents the predicted target bounding box information at the ( x , y ) point in the feature map. s x , y represents the centerness weight predicted at the ( x , y ) point in the feature map. N s represents the number of positive samples. c x , y represents the true class labels at the ( x , y ) point in the feature map. 1 s x , y > 0 is 1 when the ( x , y ) point in the feature map is matched as a positive sample, otherwise 0. 1 c x , y > 0 is 1 when the adjusted weight at the ( x , y ) point is greater than 0, otherwise 0. t x , y represents the true target bounding box information corresponding to the ( x , y ) point in the feature map. s x , y represents the weight-adjusted adaptively through the ( x , y ) point in the feature map.
Next, we transformed the depth images into point clouds and integrated them with segmented pineapple eye regions to generate corresponding pineapple eye point cloud data. We successfully stitched together point clouds for all pineapple eyes, as shown in Figure 4, achieving a high-quality reconstruction of the entire pineapple. To achieve the effect of artificially removing pineapple eyes, as shown in Figure 1b, the next step is planning the cutting path.
Next, we transformed the depth images into point clouds and integrated them with segmented pineapple eye regions to generate corresponding pineapple eye point cloud data. To facilitate understanding and computation, we establish the origin O w of the world coordinate system on the rotation axis of the pineapple, with the Z-axis aligned along the rotation axis and the Y-axis oriented vertically upward, as illustrated in Figure 2. Assuming the camera’s extrinsic matrix is denoted as T c w , and the point cloud obtained from image segmentation after each capture is represented as P c i for i { 1 , , 6 } . The world coordinate system’s point cloud P w i is given by:
P w i = w T c P c i .
After each capture, the pineapple undergoes a 60° rotation around the Z-axis, with the corresponding rotation transformation matrix denoted as T 60 . Consequently, the previously obtained world coordinate system’s point cloud P w i undergoes a rotation and is updated accordingly:
P w i = T w 60 P w i .
This rotation operation is performed each time a new set of data is collected.
Through the above operations, we successfully stitched together point clouds for all pineapple eyes, as shown in Figure 4, achieving a high-quality reconstruction of the entire pineapple. However, to achieve the effect of artificially removing pineapple eyes, as shown in Figure 1b, cutting-path planning is necessary.

2.3. Pose Correction Planning (PCP) Method

After reviewing the literature [31,32], we found that the distribution of pineapple eyes on each pineapple follows 8 helical lines. Using a specialized tool [23], we can cut along each helical line, thereby removing all pineapple eyes after cutting along the 8 helical lines.
However, during the pineapple loading process, it cannot be ensured that the rotation axis of the pineapple aligns with its own central axis. As illustrated in Figure 6, the red dashed line represents the rotational axis of the pineapple (i.e., the Z-axis of the mechanism coordinate system constructed in Section 3.1), while the green dashed line denotes the central axis of the pineapple. The challenges and focal points lie in how a computer can accurately identify the 8 helical lines of pineapple eyes and quickly and precisely remove them. To address these issues, we developed the Pose Correction Planning (PCP) method, as outlined in Algorithm 1. The key steps are as follows:
(1)
Uniform sampling of the pineapple surface point cloud ( S p ) yields a consistent and well-distributed point cloud ( S p s ) of the pineapple surface.
(2)
Use Principal Component Analysis ( m y P C A ( · ) ) to determine the pose or coordinate system ( O p ) of the pineapple.
(3)
Compute the centroid ( P w i ) of the pineapple eye surface point cloud obtained with previous sections of the paper.
(4)
Calculate the coordinates ( P p i ) of the point cloud centroid in the pineapple coordinate system.
(5)
The distribution of point cloud centroids ( P p i ) is transformed from the pineapple coordinate system to the radian-height coordinate system ( C p i ( θ z ) ), where the radian is the horizontal axis ( θ a x i s ) and the z a x i s is the vertical axis, as shown in Figure 7. It can be seen that the distribution of the pineapple eyes in this coordinate system is quite regular, with the pineapple eyes on the same spiral line essentially aligned in a straight line. Therefore, we thought of using clustering to distinguish the pineapple eyes on each spiral line.
(6)
Since the distribution radian of the pineapple eyes ranges from 0 to 2 π , we first need to shift the pineapple eyes on the same spiral line before clustering. Assuming the slope of the straight line formed by the distribution of pineapple eyes is k, and the original coordinates of pineapple eyes are C p i ( θ i , z i ) , the shifted coordinates C i ( θ i , z i ) are:
θ i = θ i 2 π if θ i z i min ( z i ) k > 2 π θ i + 2 π if θ i z i min ( z i ) k < 0 θ i otherwise .
Then, we rotate the obtained C i by α and calculate the c o s t value, and α ranges from π 6 to π 3 . The relationship between α and k is as follows:
k = 1.0 · tan ( α ) .
The coordinates after rotation, N C i , are:
N C i = cos ( α ) sin ( α ) sin ( α ) cos ( α ) · C i .
With each rotation, we perform clustering ( M y C l u s t e r A L ( · ) ) to obtain 8 groups matrices N C j ( j { 1 , , 8 } ). Each element N C i j ( θ i j , z i j ) represents a pineapple eye, where j is the index of the spiral line (ranging from 1 to 8), and i is the number of pineapple eyes on each spiral line (ranging from 1 to N u m j ). If θ i j = 0 , then N C i j ( θ i j , z i j ) can be represented as N C i j ( z i j ) . The cost function ( L c o s t ) is designed to evaluate the quality of a given scheme by penalizing the variance of different variables. Specifically:
L c o s t = k 1 V a r ( N u m j ) + k 2 1 8 j = 1 8 V a r N C i j ( z ) N C ¯ j ( z ) + k 3 1 8 j = 1 8 V a r N C [ i + 1 ] j N C [ i ] j , ( i ( 1 , 2 , , N u m j ) , j ( 1 , 2 , , 8 ) ) ,
where k 1 , k 2 and k 3 are hyperparameters that adjust the weight of each term in the overall cost function. V a r ( · ) represents the variance calculation. The calculation of N C ¯ j ( z ) is as follows:
N C ¯ j ( z ) = 1 N u m j i = 1 N u m j N C i j ( z ) .
(7)
In the mechanism coordinate system, sort the pineapple eyes on each helical line according to their Z-axis values to obtain an ordered set of pineapple eyes for each helical line. Connect the pineapple eyes on each helical line with line segments.
Algorithm 1: PCP (Pose Correction Planning) Method
Agriculture 14 01569 i001
To enhance processing efficiency, we also designed a cooperative dual-blade eye removal mechanism tailored to the specific characteristics, as illustrated in Figure 8. This mechanism comprises symmetric left and right eye-cutting devices, each equipped with three degrees of freedom, including the eye cutter, eye cutter rotary motor, eye cutter feed motor, eye cutter translation motor, and other essential components. Collaborating with the clamping platform’s pineapple rotation axis, they simultaneously cut two lines, repeat the process twice, and successfully removed the pineapple eyes along all eight spirals.

3. Results and Discussion

To validate the effectiveness of the automated processing system, we constructed a prototype of the pineapple automated processing system, as shown in Figure 8. It is comprised of an automatic feeding mechanism (including conveyor feeding and pose correction), automatic loading mechanism, peeling mechanism, RGB-D camera (D435), eye removal mechanism, and a computer (operating system: Ubuntu 20.04). This computer is equipped with 64 GB of RAM, an Intel Core i9-10940X processor, and an NVIDIA GTX3090 graphics processor. The camera has an RGB resolution of 640 × 480 pixels and operates at a frame rate of 30 fps. Different modules work in a predefined sequence, enabling the entire automated processing process. In the experiments, we conducted separate tests and statistical analyses of each module and then verified the feasibility and stability of the entire automated processing system.

3.1. Result Analysis of Pineapple Eyes Detection

In the system we constructed, the camera was installed directly above the pineapple at a distance of 35 mm from the rotation axis and calibrated with the mechanical setup. The computer is responsible for orchestrating the system’s overall control, which encompasses the regulation of individual components or subsystems, the implementation of deep vision algorithms, as well as the deployment of cutting path and trajectory planning algorithms. This visual system is a core component of our automated processing system and is responsible for identifying the position and shape of pineapple eyes.
In the actual processing environment, we processed 100 pineapples using the automatic peeling mechanism, resulting in a total of 600 pineapple images collected under varying lighting conditions. To increase data diversity and enhance the robustness of the deep vision network, we applied data augmentation techniques. Ultimately, we obtained a PineappleEyes dataset comprising 3000 pineapple images. We used the Labelme software to perform precise annotations of pineapples and pineapple eyes. The dataset was divided into training, validation, and testing sets in an 8:1:1 ratio, which were used for training, validating, and testing the deep vision network and point cloud reconstruction.
In the experiments, we compared the performance of various deep learning models, including YOLACT, SOLOv2, Mask-RCNN, BlendMask, and our enhanced SAAF-BlendMask algorithm, using the pineapple dataset established earlier. Additionally, extensive experiments were conducted to explore different values of the hyperparameters m and n. The experimental results, as depicted in Table 1 and Table 2, highlight the exceptional performance of our enhanced algorithm. The detection results for a single instance of pineapple eye recognition are shown in Figure 9.
To further analyze the impact of the PineappleEyes dataset splitting on model stability, we conducted additional experiments with a different data split ratio of 6:2:2 (training:validation:testing). These experiments were performed using the same models: YOLACT, SOLOv2, Mask-RCNN, BlendMask, and SAAF-BlendMask. As presented in Table 3, the results demonstrate that the SAAF-BlendMask algorithm maintains its superior performance across different dataset splits. Specifically, under the 6:2:2 split, SAAF-BlendMask continued to outperform other algorithms in terms of both average precision (AP) and small object precision (APs).
It is evident that the inference time of the BlendMask algorithm and our improved BlendMask algorithm are closely comparable, significantly outperforming the Mask-RCNN algorithm by approximately 62% in inference time. Moreover, compared to the Mask-RCNN algorithm, our enhanced BlendMask algorithm demonstrated a higher AP50. As shown in Table 2, when m = 1 and n = 30 , the A P for Mask-RCNN improved by nearly 0.2 percentage points, and for BlendMask it improved by nearly 2.78 percentage points, reaching 73.04%. For the recognition metric focused on small objects, A P s , the enhanced BlendMask algorithm outperformed BlendMask by nearly 6 percentage points, reaching 62.54%. Under the 6:2:2 split, as presented in Table 3, the trends were consistent with those observed in the 8:1:1 split. The A P was also nearly 7.6 percentage points higher than BlendMask, reaching 74.70%, while the APs improved by almost 9 percentage points, reaching 54.78%, further reinforcing the robustness and adaptability of the proposed SAAF-BlendMask algorithm.
In addition, to further validate the performance and interpretability of the SAAF-BlendMask algorithm, we conducted an additional set of experiments using a subset of the GraspNet-1Billion dataset [33]. This dataset is widely recognized for containing small objects, making it an appropriate benchmark to assess the algorithm’s generalization capability beyond the specific domain of pineapple processing. Given the substantial size of the GraspNet-1Billion dataset, we selected one image from each of 192 different scenes to create a smaller dataset for our testing. This approach allowed us to efficiently evaluate the algorithm while ensuring that the subset was representative of the dataset’s overall diversity. The results of this comparative study are presented in Table 4. The SAAF-BlendMask algorithm achieved the highest AP score of 85.53%, outperforming other methods across several metrics, particularly in detecting small objects (APs) with a score of 29.34%.
These results indicate that the SAAF-BlendMask algorithm is not only effective in the specific context of pineapple processing but also demonstrates strong generalization capabilities when applied to external datasets with different characteristics. Despite using a smaller subset of the GraspNet-1Billion dataset, the algorithm’s performance remained robust, further validating its potential for broader applications.
Following pineapple eye segmentation on the RGB images and alignment with the depth images, we extract the corresponding pineapple eye regions from the depth images and convert them into point clouds. During the processing workflow, we capture six images uniformly for each pineapple. We repeat the aforementioned procedure, performing point cloud fusion, denoising, and filtering operations to reconstruct a complete pineapple eye point cloud in the equipment coordinate system, as depicted in Figure 4.

3.2. Result Analysis of Pineapple Eyes Cutting Path Planning

To validate the effectiveness of the PCP method, we conducted a comparative experiment using 100 pineapples. After peeling the pineapples with a machine and reconstructing the pineapple eye point clouds, we planned the cutting path. Initially, the PCP method was not applied, and the results, as shown in Figure 10a,b, depict the distribution of pineapple eyes in the machine coordinate system. Subsequently, the same pineapples underwent cutting path planning using the PCP method, with the pineapple eyes distributed along eight helical lines in the pineapple’s coordinate system, as illustrated in Figure 10c,d. It is clear that the path planned using the PCP method perfectly forms eight spiral lines, achieving the desired effect. This method is more suitable for dual-blade coordinated cutting. The results clearly indicate that path planning with the PCP method is more reliable. The experimental outcomes reveal that 75 pineapples were successfully planned without the PCP method, whereas 99 pineapples were successfully planned with the PCP method.
Following the distribution of pineapple eyes along each spiral, we proceed to plan the cutting trajectories. As outlined in the design of the eye removal apparatus in the preceding sections, we understand that during eye removal, the left and right symmetric eye removal mechanisms collectively exhibit six axes that coordinate with the pineapple’s rotation axis. Using the pineapple rotation axis as the reference, we employ cubic spline interpolation three times for the remaining six axes to obtain the motion trajectories or interpolation data for the eye removal.
During the eye removal process, the computer dynamically controls the synchronized operation of the symmetrical eye removal mechanisms in real-time based on interpolation data, achieving pineapple eye removal (as shown in Figure 11) in a manner that emulates human dexterity (as shown in Figure 1b). In the experiments, we selected 110 pineapples for cutting, divided into four groups based on weight. Initially, we conducted a cutting experiment using our method. After weighing, the peeled pineapples were further processed to achieve the same peeling and eye removal results as traditional methods. The experimental data are presented in Table 5. Our method of pineapple peeling has demonstrated a significant advantage over the traditional method in terms of minimizing waste. Specifically, the total peeling waste using our method was 40.78 kg, compared to 46.18 kg with traditional automatic processing methods such as those in [3]. The processing effects are similar to those shown in Figure 1a. This indicates that our method preserved approximately 5.40 kg more pineapple flesh, representing a reduction in waste of about 11.7%.
While automatic processing systems can process up to 75 pineapples per minute [3], our approach focuses on optimizing the balance between speed and waste reduction. Traditional manual methods require around 2 min per pineapple [2], but our method achieves the same precision in just 30 s. This balance of efficiency and precision demonstrates the strength of our system in applications where minimizing waste is a priority, even though it may not match the speed of fully automated systems.

3.3. Comparison of Different Pineapple Processing Methods

We also conducted extensive research in the specific domain of automated pineapple processing. Table 6 compares some of the most popular reported processing methods. Clearly, through comparison, the proposed approach demonstrates more advantages than all methods. The column Integrated System addresses the question of whether a fully automated electromechanical integrated system is provided to address this issue. This paper also compares the use of machine learning model integration. It can be inferred that the proposed system not only fairly addresses the initial problem but also leverages and promotes the use of cutting-edge technology to ensure its relevance in the near future.

4. Conclusions

The developed algorithms, particularly the Scale-Adaptive Adjustment Factor-enhanced BlendMask algorithm and the PCP method, have shown exceptional effectiveness in tackling the challenges of precision pineapple processing. Rigorous experimentation and analysis have demonstrated that these algorithms outperform conventional methods, especially in the areas of pineapple eye detection and path planning accuracy.
The integration of these advanced algorithms into an automated processing system marks a significant advancement in pineapple processing technology. The system, equipped with 3D vision and deep learning capabilities, offers a robust solution for enhancing processing efficiency and reducing fruit wastage in industrial applications.
While this research primarily focuses on the technical aspects of pineapple processing, such as automation and efficiency, it is important to note that our system is designed with minimal human intervention, which inherently enhances food safety and processing consistency. However, the study did not delve deeply into post-harvest quality attributes such as color, texture, tissue damage, or sugar content. These factors are crucial in evaluating the overall quality of processed pineapples, especially for fresh consumption. Our processing method is mainly intended for canned pineapple production, where the primary concern is maximizing yield and minimizing waste rather than optimizing fresh quality attributes. Future work could explore these quality factors in greater detail to provide a more comprehensive comparison between traditional and automated processing methods.
The experimental findings affirm the system’s effectiveness, emphasizing its robustness, reliability, and adaptability in practical scenarios. By optimizing processing workflows and minimizing waste, the system not only boosts productivity but also contributes to resource conservation and sustainability within the food industry.
In summary, the combination of these innovative algorithms and advanced system development represents a transformative approach to pineapple processing. By leveraging cutting-edge technologies, this research sets the stage for more efficient, sustainable, and economically viable agricultural product processing systems.

Author Contributions

Conceptualization, H.W., H.Z., and J.D.; methodology, H.W., H.Z., and J.D.; software, H.Z., C.L., and Y.Z.; validation, H.W., J.D., and J.T.; formal analysis, Y.Z., and J.T.; investigation, H.W.; resources, J.T.; data curation, Y.Z. and C.L.; writing—original draft preparation, H.W.; writing—review and editing, H.Z. and C.L.; visualization, Y.Z. and J.D.; supervision, C.L.; project administration, J.T.; funding acquisition, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science and Technology Service Network Initiative (STS) Project of Chinese Academy of Sciences, grant number STS-HP-202207.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All data are presented in this article in the form of figures and tables.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mohsin, A.; Jabeen, A.; Majid, D.; Allai, F.M.; Dar, A.H.; Gulzar, B.; Makroo, H.A. Pineapple. In Antioxidants in Fruits: Properties and Health Benefits; Springer: Berlin/Heidelberg, Germany, 2020; pp. 379–396. [Google Scholar]
  2. Rohrbach, K.G.; Leal, F.; d’Eeckenbrugge, G.C. History, distribution and world production. In The Pineapple: Botany, Production and Uses; CABI Digital Library: Oxfordshire, UK, 2003; pp. 1–12. [Google Scholar]
  3. Bartholomew, D.P.; Hawkins, R.A.; Lopez., J.A. Hawaii pineapple: The rise and fall of an industry. HortScience 2012, 47, 1390–1398. [Google Scholar] [CrossRef]
  4. Jia, W.; Zhang, Z.; Shao, W.; Hou, S.; Ji, Z.; Liu, G.; Yin, X. Foveamask: A fast and accurate deep learning model for green fruit instance segmentation. Comput. Electron. Agric. 2021, 191, 106488. [Google Scholar] [CrossRef]
  5. Kasinathan, T.; Singaraju, D.; Uyyala, S.R. Insect classification and detection in field crops using modern machine learning techniques. Inf. Process. Agric. 2021, 8, 446–457. [Google Scholar] [CrossRef]
  6. Koirala, A.; Walsh, K.B.; Wang, Z.; McCarthy, C. Deep learning—Method overview and review of use for fruit detection and yield estimation. Comput. Electron. Agric. 2019, 162, 219–234. [Google Scholar] [CrossRef]
  7. Indira, D.; Goddu, J.; Indraja, B.; Challa, V.M.L.; Manasa, B.A. review on fruit recognition and feature evaluation using cnn. Mater. Today Proc. 2023, 80, 3438–3443. [Google Scholar] [CrossRef]
  8. Shakya, D.S. Analysis of artificial intelligence based image classifica- tion techniques. J. Innov. Image Process. 2020, 2, 44–54. [Google Scholar] [CrossRef]
  9. Zhang, W.; Zhao, D.; Gong, W.; Li, Z.; Lu, Q.; Yang, S. Food image recognition with convolutional neural networks. In Proceedings of the 2015 IEEE 12th International Conference on Ubiquitous Intelligence and Computing and 2015 IEEE 12th International Conference on Autonomic and Trusted Computing and 2015 IEEE 15th International Conference on Scalable Computing and Communications and Its Associated Workshops (UIC-ATC-ScalCom), Beijing, China, 10–14 August 2015; pp. 690–693. [Google Scholar]
  10. Matsuda, Y.; Hoashi, H.; Yanai, K. Recognition of multiple-food images by detecting candidate regions. In Proceedings of the 2012 IEEE International Conference on Multimedia and Expo, Melbourne, Australia, 9–13 July 2012; pp. 25–30. [Google Scholar]
  11. Liu, G.; Nouaze, J.C.; Touko Mbouembe, P.L.; Kim, J.H. Yolo-tomato: A robust algorithm for tomato detection based on yolov3. Sensors 2020, 20, 2145. [Google Scholar] [CrossRef]
  12. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Yu, Y.; Zhang, K.; Yang, L.; Zhang, D. Fruit detection for strawberry harvesting robot in non-structural environment based on mask-rcnn. Comput. Electron. Agric. 2019, 163, 104846. [Google Scholar] [CrossRef]
  14. He, K.; Gkioxari, G.; Doll’ar, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  15. Wang, X.A.; Tang, J.; Whitty, M. Deepphenology: Estimation of apple flower phenology distributions based on deep learning. Comput. Electron. Agric. 2021, 185, 106123. [Google Scholar] [CrossRef]
  16. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  17. Chen, S.W.; Shivakumar, S.S.; Dcunha, S.; Das, J.; Okon, E.; Qu, C.; Taylor, C.J.; Kumar, V. Counting apples and oranges with deep learning: A data-driven approach. IEEE Robot. Autom. Lett. 2017, 2, 781–788. [Google Scholar] [CrossRef]
  18. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  19. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  20. Haris, M.; Shakhnarovich, G.; Ukita, N. Task-driven super resolution: Object detection in low-resolution images. In Proceedings of the 28th International Conference on Neural Information Processing, Sanur, Bali, Indonesia, 8–12 December 2021; pp. 387–395. [Google Scholar]
  21. Bai, Y.; Zhang, Y.; Ding, M.; Ghanem, B. Sod-mtgan: Small object detection via multi-task generative adversarial network. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 206–221. [Google Scholar]
  22. Lin, T.-Y.; Doll’ar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  23. Singh, V.; Verma, D.K.; Singh, G. Development of pineapple peeler- cum-slicer. Pop. Kheti 2013, 1, 21–24. [Google Scholar]
  24. Kumar, P.; Chakrabarti, D.; Patel, T.; Chowdhuri, A. Work-related pains among the workers associated with pineapple peeling in small fruit processing units of north east india. Int. J. Ind. Ergon. 2016, 53, 124–129. [Google Scholar] [CrossRef]
  25. Anjali, A.; Anjitha, P.; Neethu, K.; Arunkrishnan, A.; Vahid, P.A.; Mathew, S.M. Development and performance evaluation of a semi- automatic pineapple peeling machine. Int. J. Curr. Appl. Sci. 2019, 8, 325–332. [Google Scholar] [CrossRef]
  26. Jongyingcharoen, J.S.; Cheevitsopon, E. Design and development of continuous pineapple-peeling machine. Agric. Nat. 2022, 56, 979–986. [Google Scholar]
  27. Siriwardhana, P.G.A.L.; Wijewardane, D.C. Machine for the pineapple peeling process. J. Eng. Technol. Open Univ. Sri Lanka (JET-OUSL) 2018, 6, 1–15. [Google Scholar]
  28. Kumar, P.; Chakrabartiand, D. Design of pineapple peeling equipment. In Proceedings of the IEEE 6th International Conference on Advanced Production and Industrial Engineering, Delhi, India, 18–19 June 2021; pp. 545–556. [Google Scholar]
  29. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one- stage object detection. In Proceedings of the International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  30. Chen, H.; Sun, K.; Tian, Z.; Shen, C.; Huang, Y.; Yan, Y. Blendmask: Top-down meets bottom-up for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8573–8581. [Google Scholar]
  31. Ekern, P.C. Phyllotaxy of pineapple plant and fruit. Bot. Gaz. 1968, 129, 92–94. [Google Scholar] [CrossRef]
  32. Hikal, W.M.; Mahmoud, A.A.; Said-Al Ahl, H.A.; Bratovcic, A.; Tkachenko, K.G.; Kačániová, M.; Rodriguez, R.M. Pineapple (ananas comosus l. merr.), waste streams, characterisation and valorisation: An overview. Open J. Ecol. 2021, 11, 610–634. [Google Scholar] [CrossRef]
  33. Fang, H.-S.; Wang, C.; Gou, M.; Lu, C. Graspnet-1billion: A large-scale benchmark for general object grasping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11441–11450. [Google Scholar]
  34. Liu, A.; Xiang, Y.; Li, Y.; Hu, Z.; Dai, X.; Lei, X.; Tang, Z. 3D positioning method for pineapple eyes based on multiangle image stereo-matchig. Agriculture 2022, 12, 2039. [Google Scholar] [CrossRef]
Figure 1. Pineapple processed in the traditional way. (a) Pineapple processed in an automated manner, like the Ginaca machine [2,3]. (b) Pineapple is processed by hand.
Figure 1. Pineapple processed in the traditional way. (a) Pineapple processed in an automated manner, like the Ginaca machine [2,3]. (b) Pineapple is processed by hand.
Agriculture 14 01569 g001
Figure 2. The processing system. Mainly consists of the clamping platform mechanism, peeling mechanism, vision system, and eye removal mechanism.
Figure 2. The processing system. Mainly consists of the clamping platform mechanism, peeling mechanism, vision system, and eye removal mechanism.
Agriculture 14 01569 g002
Figure 3. RGB-D camera calibration. (a) The field of view for the D435 depth camera is depicted as a frustum. The central region of the frustum represents the acceptable distance of the surface from the camera coordinate system, (b) A schematic diagram illustrates the coordinate transformation relationship between world, camera, and pixel frames. R and t denote the rotation and translation elements constituting the homogeneous transform for extrinsic parameters, while K represents the matrix embodying intrinsic parameters.
Figure 3. RGB-D camera calibration. (a) The field of view for the D435 depth camera is depicted as a frustum. The central region of the frustum represents the acceptable distance of the surface from the camera coordinate system, (b) A schematic diagram illustrates the coordinate transformation relationship between world, camera, and pixel frames. R and t denote the rotation and translation elements constituting the homogeneous transform for extrinsic parameters, while K represents the matrix embodying intrinsic parameters.
Agriculture 14 01569 g003
Figure 4. The overall structure of the vision system algorithm. It comprises two main components: instance segmentation and 3D point cloud reconstruction. The instance segmentation, based on the improved BlendMask [30], is further divided into four major modules: the feature extraction module (denoted by the blue dashed box), object detection module (represented by the orange dashed box), feature prediction module (indicated by the green dashed box), and fusion module (depicted by the orange-green block). This constructed network facilitates the rapid segmentation of pineapple and pineapple eyes’ outlines, thereby enhancing the recognition rate of small-target pineapple eyes. The point cloud reconstruction section includes depth image processing, point cloud filtering, and pineapple eye reconstruction.
Figure 4. The overall structure of the vision system algorithm. It comprises two main components: instance segmentation and 3D point cloud reconstruction. The instance segmentation, based on the improved BlendMask [30], is further divided into four major modules: the feature extraction module (denoted by the blue dashed box), object detection module (represented by the orange dashed box), feature prediction module (indicated by the green dashed box), and fusion module (depicted by the orange-green block). This constructed network facilitates the rapid segmentation of pineapple and pineapple eyes’ outlines, thereby enhancing the recognition rate of small-target pineapple eyes. The point cloud reconstruction section includes depth image processing, point cloud filtering, and pineapple eye reconstruction.
Agriculture 14 01569 g004
Figure 5. The performance of FCOS. (a) The performance of anchor point on large target. (b) The performance of anchor point on small target. In the diagram, the rectangles represent the ground truth (GT) boxes predicted by FCOS, the blue dots represent anchor points, and l , t , r , and b denote the distances from the anchor points to the left, top, right, and bottom sides of the boxes, respectively.
Figure 5. The performance of FCOS. (a) The performance of anchor point on large target. (b) The performance of anchor point on small target. In the diagram, the rectangles represent the ground truth (GT) boxes predicted by FCOS, the blue dots represent anchor points, and l , t , r , and b denote the distances from the anchor points to the left, top, right, and bottom sides of the boxes, respectively.
Agriculture 14 01569 g005
Figure 6. The clamping state of the pineapple after feeding and peeling. The red dashed line represents the rotational axis of the pineapple, and the green dashed line denotes the central axis of the pineapple.
Figure 6. The clamping state of the pineapple after feeding and peeling. The red dashed line represents the rotational axis of the pineapple, and the green dashed line denotes the central axis of the pineapple.
Agriculture 14 01569 g006
Figure 7. Display of pineapple eyes in the θ z coordinate system.
Figure 7. Display of pineapple eyes in the θ z coordinate system.
Agriculture 14 01569 g007
Figure 8. The vision-based pineapple automated processing system we constructed.
Figure 8. The vision-based pineapple automated processing system we constructed.
Agriculture 14 01569 g008
Figure 9. Detection results of pineapple eyes using our algorithm. The images demonstrate the identification performance of our proposed algorithm.
Figure 9. Detection results of pineapple eyes using our algorithm. The images demonstrate the identification performance of our proposed algorithm.
Agriculture 14 01569 g009
Figure 10. Pineapple eye-cutting path-planning results. The cutting path consists of 8 lines covering the pineapple’s surface, and each point represents the centroid of the pineapple eye point cloud with numbering. (a) Path planning results for pineapple eye cutting in the θ z coordinate system without using the posture correction algorithm. (b) Path planning results for pineapple eye cutting in the Cartesian coordinate system without using the posture correction algorithm. (c) Path planning results for pineapple eye cutting in the θ z coordinate system with using the posture correction algorithm. (d) Path planning results for pineapple eye cutting in the Cartesian coordinate system with using the posture correction algorithm. In (a,c), the points represent the centroids of different pineapple eyes. The eight different colors of the points indicate the results of dividing the pineapple eyes onto different spirals (represented as line segments in the θ z coordinate system) after applying the path planning algorithm. In (b,d), the blue points represent the centroids of the pineapple eyes, and the different colored polylines indicate the effects of dividing the pineapple eyes onto different spirals in the Cartesian coordinate system after applying the path planning algorithm. (a,b) correspond to the results of using only the path planning algorithm, while (c,d) show the results of using the PCP path planning algorithm. The correspondence between the pineapple eyes can be viewed by their numbering.
Figure 10. Pineapple eye-cutting path-planning results. The cutting path consists of 8 lines covering the pineapple’s surface, and each point represents the centroid of the pineapple eye point cloud with numbering. (a) Path planning results for pineapple eye cutting in the θ z coordinate system without using the posture correction algorithm. (b) Path planning results for pineapple eye cutting in the Cartesian coordinate system without using the posture correction algorithm. (c) Path planning results for pineapple eye cutting in the θ z coordinate system with using the posture correction algorithm. (d) Path planning results for pineapple eye cutting in the Cartesian coordinate system with using the posture correction algorithm. In (a,c), the points represent the centroids of different pineapple eyes. The eight different colors of the points indicate the results of dividing the pineapple eyes onto different spirals (represented as line segments in the θ z coordinate system) after applying the path planning algorithm. In (b,d), the blue points represent the centroids of the pineapple eyes, and the different colored polylines indicate the effects of dividing the pineapple eyes onto different spirals in the Cartesian coordinate system after applying the path planning algorithm. (a,b) correspond to the results of using only the path planning algorithm, while (c,d) show the results of using the PCP path planning algorithm. The correspondence between the pineapple eyes can be viewed by their numbering.
Agriculture 14 01569 g010
Figure 11. The resulting pineapple after processing with our method. The images demonstrate the performance of our proposed algorithm in removing pineapple eyes.
Figure 11. The resulting pineapple after processing with our method. The images demonstrate the performance of our proposed algorithm in removing pineapple eyes.
Agriculture 14 01569 g011
Table 1. Evaluation results for segmentation (SEGM) with SAAF-BlendMask.
Table 1. Evaluation results for segmentation (SEGM) with SAAF-BlendMask.
mnAPAP50AP75APsAPlInfer Time (ms)
12072.3494.1782.3162.0583.190.020311
102072.3394.4880.6860.7283.940.020599
202074.2794.9582.4762.4486.110.020481
302073.3794.9682.7161.8284.920.01985
13073.0495.4483.6262.5483.560.020537
103072.1092.2582.8762.3582.250.020342
203070.8492.7881.7462.6979.000.020744
303070.5791.8681.2962.5678.240.020971
Table 2. Evaluation results for segmentation (SEGM) with different methods on 8:1:1 PineappleEyes Dataset Split.
Table 2. Evaluation results for segmentation (SEGM) with different methods on 8:1:1 PineappleEyes Dataset Split.
MethodsAPAP50AP75APsAPlInfer Time (ms)
YOLACT72.8994.0282.7860.4284.510.014947
SOLOv272.1193.1780.8658.4484.720.014527
Mask-RCNN72.8592.6981.7762.2983.420.053034
BlendMask70.2694.4177.5956.5583.960.019867
SAAF-BlendMask (Ours)73.0495.4483.6262.5483.560.020537
Table 3. Evaluation results for segmentation (SEGM) with different methods on 6:2:2 PineappleEyes Dataset Split.
Table 3. Evaluation results for segmentation (SEGM) with different methods on 6:2:2 PineappleEyes Dataset Split.
MethodsAPAP50AP75APsAPlInfer Time (ms)
YOLACT70.8697.4669.0046.9594.780.014475
SOLOv269.4693.3667.7442.6097.330.015145
Mask-RCNN70.9698.7079.4054.2586.780.054751
BlendMask67.0597.0765.5045.9291.180.019354
SAAF-BlendMask (Ours)74.7098.6576.9354.7895.610.020245
Table 4. Evaluation results for segmentation (SEGM) with different methods on the GraspNet-1Biliion Dataset.
Table 4. Evaluation results for segmentation (SEGM) with different methods on the GraspNet-1Biliion Dataset.
MethodsAPAP50AP75APsAPlInfer Time (ms)
YOLACT73.8287.3979.0624.1974.520.014647
SOLOv285.0192.3891.5325.9289.160.015024
Mask-RCNN78.8395.2187.5118.5280.690.056247
BlendMask85.2492.3891.7622.1387.810.019875
SAAF-BlendMask(Ours)85.5392.5089.5729.3488.020.020247
Table 5. Data on pineapple cutting and peeling with different methods.
Table 5. Data on pineapple cutting and peeling with different methods.
Weight Group (kg)<1.21.2–1.71.7–2.2>2.2Total
Number of Pineapples2262197110
Total Weight (kg)22.6592.5137.4117.16169.73
Peeling Waste with
Our Method (kg)
5.6822.348.793.9740.78
Peeling Waste with Traditional
Automatic Processing Method [3] (kg)
6.3424.9210.144.7846.18
This table shows the data on pineapple cutting and peeling using different methods. The waste data are calculated based on different peeling techniques.
Table 6. Comparative analysis of the proposed framework with the most mainstream techniques.
Table 6. Comparative analysis of the proposed framework with the most mainstream techniques.
MethodsISLLCLFWEPAIMEMCost
Full Manual
Processing [23,25]
×××××HLC, LEC
Automatic
Processing [3]
××HEC
Semi-auto
Peeling [26]
×××××MEC, MLC
Spiral Peeling
Processing [27]
×××××MEC, MLC
Visual Recognition
Pineapple Eyes [34]
×××××MEC, HLC
OursMEC
“✓” indicates that the method possesses the corresponding attribute, while “×” indicates that the method does not. Abbreviations: IS (Integrated System), LLC (Low Labor Costs), LFW (Low FleshWastage), EP (Efficient Processing), AIM (AI-based Method), EM (Ensemble Methods), HLC (High Labor Costs), HEC (High Equipment Costs), MEC (Moderate Equipment Costs), MLC (Moderate Labor Costs), LEC (Low Equipment Costs).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Zhang, H.; Zhang, Y.; Deng, J.; Liu, C.; Tan, J. Integrated Scale-Adaptive Adjustment Factor-Enhanced BlendMask Method for Pineapple Processing System. Agriculture 2024, 14, 1569. https://doi.org/10.3390/agriculture14091569

AMA Style

Wang H, Zhang H, Zhang Y, Deng J, Liu C, Tan J. Integrated Scale-Adaptive Adjustment Factor-Enhanced BlendMask Method for Pineapple Processing System. Agriculture. 2024; 14(9):1569. https://doi.org/10.3390/agriculture14091569

Chicago/Turabian Style

Wang, Haotian, Haojian Zhang, Yukai Zhang, Jieren Deng, Chengbao Liu, and Jie Tan. 2024. "Integrated Scale-Adaptive Adjustment Factor-Enhanced BlendMask Method for Pineapple Processing System" Agriculture 14, no. 9: 1569. https://doi.org/10.3390/agriculture14091569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop