Skip to Content
AgricultureAgriculture
  • Article
  • Open Access

25 July 2024

Flexible Hand Claw Picking Method for Citrus-Picking Robot Based on Target Fruit Recognition

,
,
and
1
College of Electrical and Information Engineering, Hunan University, Changsha 410082, China
2
National Engineering Research Center for Robot Vision Perception and Control Technology, Hunan University, Changsha 410082, China
3
College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Agricultural Technology

Abstract

In order to meet the demand of the intelligent and efficient picking of fresh citrus fruit in a natural environment, a flexible and independent picking method of fresh citrus fruit based on picking pattern recognition was proposed. The convolutional attention (CA) mechanism was added in the YOLOv7 network model. This makes the model pay more attention to the citrus fruit region, reduces the interference of some redundant information in the background and feature maps, effectively improves the recognition accuracy of the YOLOv7 network model, and reduces the detection error of the hand region. According to the physical parameters of the citrus fruit and stem, an end-effector suitable for picking citrus fruit was designed, which effectively reduced the damage during the picking of citrus fruit. According to the actual distribution of citrus fruits in the natural environment, a citrus fruit-picking task planning model was established, so that the adaptability of the flexible handle can make up for the inaccuracy of the deep learning method to a certain extent when the end-effector picks fruits independently. Finally, on the basis of integrating the key components of the picking robot, a production test was carried out in a standard citrus orchard. The experimental results show that the success rate of the citrus-picking robot arm is 87.15%, and the success rate of picking in the natural field environment is 82.4%, which is better than the success rate of 80% of the market picking robot. In the picking experiment, the main reason for the unsuccessful positioning of citrus fruits is that the position of citrus fruits is beyond the picking range of the end-effector, and the motion parameters of the robot arm joint will produce errors, affecting the motion accuracy of the robot arm, leading to the failure of picking. This study can provide technical support for the exploration and application of the intelligent fruit-picking mode.

1. Introduction

Fruit is an indispensable part of the human diet and a significant source of income through cash crops [1,2]. Data from the National Bureau of Statistics indicate that over the past decade, China’s fruit cultivation area and output have been consistently increasing. By 2021, these figures were projected to reach 12,962 square kilometers and 296.11 million tons, respectively, placing China at the forefront globally [3,4]. Since 2004, advancements in agricultural mechanization have enabled the complete mechanization of traditional grain crops, such as rice and wheat, from sowing to harvesting. However, fruit harvesting remains largely dependent on manual labor, which is labor-intensive [5].
At present, researchers in China and abroad have conducted a lot of research on the identification of fruit trees. The main method of fruit recognition is visual recognition, that is, image recognition, which uses image acquisition equipment to collect images, and then classifies and recognizes the images. Traditional image recognition methods are mainly used to extract the features from the image manually, and then to recognize the fruit. With the development of computer technology, image recognition based on machine learning methods and deep convolutional neural networks has been widely used in fruit detection. (1) Traditional visual fruit recognition: RGB images captured by image acquisition devices include major features such as color, shape, and texture. Early studies have generally focused on extracting a single feature to identify fruit. Qian Jianping et al. introduced the mixed color space recognition method with a V value on the basis of RGB color, which effectively improved the recognition success rate of apples under different lighting conditions [6]. (2) Fruit recognition based on machine learning methods: the above artificial feature extraction methods are time-consuming and laborious, and a large number of feature combination experiments are needed to obtain the best results. Since the 1980s, machine learning has developed rapidly. Many scholars combine machine learning theory with fruit recognition technology, and mainly use SVM, Canny, HOG, and other methods for feature extraction to enable fruit recognition. Wei et al. proposed a fruit recognition method using color features based on the improved OSTU threshold algorithm and OHTA color feature model, and a large number of tests proved that the recognition accuracy of this recognition method was above 95% [7]. Arefi et al. used the combination of RGB, HSI, and YIQ space to extract the color features of ripe tomatoes, and the success rate of identifying tomato fruits in a greenhouse was about 96.36% [8]. (3) Fruit recognition based on convolutional neural networks: Ref. [9] used RGB and YCbCr to perform color segmentation on the acquired images and performed variable texture segmentation on adjacent pixels in the image, so as to enable the detection and recognition of mango. This recognition method can effectively reduce the influence of light on the recognition effect. Ref. [10] optimized the structure of the convolution layer and the pooling layer in a Faster R-CNN model, making the identification more accurate and faster. The ability to recognize fruits in different stages of maturity and with complex backgrounds, and the ability to generalize in actual orchard environments, need to be further optimized with algorithms and adjusted model parameters. Compared with the traditional recognition methods, algorithms have higher detection accuracy and shorter processing times, and the average accuracy of detecting multiple fruits on a homemade fruit dataset reaches 91% [11].
In the early 1980s, Japanese scholars initiated research on fruit-picking robots, with significant contributions from experts such as Kondo and Kadota Charishi, whose findings have substantially influenced the development of this technology. Recently, Yukio Honda of Panasonic has spearheaded the development of a greenhouse tomato-picking robot capable of gently harvesting fruit and transporting it to a cart, as well as automatically replacing the harvest box. This innovation aims to reduce daytime labor by enabling night-time automated picking. Researchers at Kochi University of Technology, including Bachche et al. [12], have developed an articulated arm robot designed for harvesting bell peppers, particularly those grown in V-frame structures within greenhouses. Yaguchi et al. [13] from the University of Tokyo have utilized advanced technology, including an electric wheeled omni-directional chassis, a UR5 6-axis robotic arm, and a Sony PS4 binocular stereo camera, coupled with a gripper-twist end-effector, to create a tomato-picking robot. This robot is capable of operating in natural light within narrow greenhouse aisles. However, challenges remain, including issues with grip failure, damage to the fruit’s calyx, and difficulties in gripping multiple fruits simultaneously. Chen et al. [14] from the University of Tokyo have also made strides with their development of a humanoid two-armed tomato-picking robot. This robot features an omni-directional chassis and is equipped with Xtion and Carmine somatosensory cameras on its head and wrist, respectively. The picking arm boasts seven degrees of freedom and a scissor-type end-effector. While the robot has been tested for indoor hanging tomato-picking, it currently requires human command input to perform its tasks. Further enhancements are needed in the areas of identification, localization, and the picking process. In the United States, the Energid company [15] has received financial backing from the U.S. Department of Agriculture to develop a citrus-picking robot. This robot is mounted on a diesel truck and uses visual sensors and high-speed computers for information acquisition, processing, analysis, and control. According to the company’s website, the robot can pick each citrus fruit in 2–3 s, with an approximate picking rate of 50%.
Domestic research into fruit-picking robots commenced in the mid-1990s. At Jiangsu University, Refs. [16,17] developed a prototype mobile tomato-picking robot equipped with a wheeled chassis, a clamping and shearing end-effector, and a binocular vision system. This prototype, along with greenhouse fruit-picking and transporting robots, was designed to work in coordinated operations to achieve full automation in tomato-picking, on-site grading, collection, transportation, and unloading. Refs. [18,19] from the same university developed an apple-picking robot capable of automatically detecting and locating apples in trees using a support vector machine with radial basis functions. This robot is equipped with a five-degree-of-freedom PRRRP-type robotic arm to autonomously perform the harvesting task. Researchers at Nanjing Agricultural University, including Ref. [20], have developed an intelligent mobile fruit-picking robot designed for low and dense planting type orchards. Outdoor picking experiments have been conducted using apple fruit trees, demonstrating the robot’s capabilities in autonomous navigation, robotic arm motion control, fruit identification and localization, end-effector fruit grasping, and fruit boxing. The National Agricultural Intelligent Equipment Engineering Technology Research Center, with Feng Youth and colleagues [21], has developed a tomato-picking robot for hanging line cultivation. This robot features a rail-type mobile lifting platform, a four-degree-of-freedom articulated robotic arm, and a line laser vision system. It uses a CCD camera and laser vertical scanning for fruit identification and localization. Experiments showed that the robot could pick a single tomato fruit in about 24 s, with a picking success rate of 83.9% under strong light and 79.4% under low light. The team led by Professor Ref. [22] has developed a kiwifruit-picking robot that uses a right-angled coordinate robotic arm to pick the fruit from underneath. The average time to pick a single fruit is 11.61 s, with a success rate of 91.7%. Refs. [23,24] proposed a two-armed tomato-picking robot, equipped with two three-degree-of-freedom robotic arms and two different types of end-effectors. These arms can cooperate to harvest tomatoes, thereby improving the harvesting efficiency [25,26,27].
In order to solve the above problems, and in order to meet the demand of the intelligent and efficient picking of fresh citrus fruit in natural environment, a flexible and independent picking method of fresh citrus fruit based on picking pattern recognition was proposed. The CA mechanism was added in the YOLOv7 network model. This makes the model pay more attention to the citrus fruit region, reduces the interference of some redundant information in the background and feature maps, effectively improves the recognition accuracy of the YOLOv7 network model, and reduces the detection error of the hand region. According to the physical parameters of the citrus fruit and stem, an end-effector suitable for picking citrus fruit was designed, which effectively reduced the damage during the picking of citrus fruit. According to the actual distribution of citrus fruits in the natural environment, a citrus fruit-picking task planning model was established, so that the adaptability of the flexible handle could make up for the inaccuracy of the deep learning method to some extent when the end-effector picked fruits independently. Finally, on the basis of integrating the key components of the picking robot, the key performance of the citrus-picking robot is verified and analyzed, which provides a reference for the research and development and application of orchard intelligent picking equipment.

3. Experimental Results and Analysis

3.1. Improved YOLOv7 Model for Fruit Recognition Results

In order to demonstrate the effectiveness of the model improvement, the detection networks of different frameworks are comparatively analyzed, and the detection results in the complete test set are shown in Table 2. From the table, it can be seen that the improved YOLOv7 model proposed in this paper has a detection accuracy P, recall R, reconciliation mean F1, and average precision AP of 95.27%, 93.28%, 92.88%, and 96.55%, respectively, and compared with the classical detection Faster RCNN and SSD models, the F1 value of the improved method proposed in this paper has been improved by 12.22 and 10.65 percentage points, and the average detection accuracy is improved by 2.69 and 3.11 percentage points, respectively. Compared with the traditional YOLOv7 model, the improved model also has some advantages in detection accuracy, and the average accuracy AP values are improved by 0.92 and 0.72 percentage points, respectively.
Table 2. Detection results of different detection network models.
Considering the detection accuracy and real-time and lightweight degree, the improved YOLOv7 model proposed in this paper shows the best comprehensive detection performance, which can accurately obtain the fruit position information on the tree in real time and meet the requirements of the picking operation. The loss value l of the model training process is shown in Figure 11, and the model reaches the convergence state in the 71st generation.
Figure 11. Loss value change curve during training.

3.2. Citrus Fruit Positioning Results

The real-time detection of citrus fruits was carried out in the citrus orchard of the Hunan Academy of Agricultural Sciences, where the depth camera was fixed on a tripod to test the recognition and localization of citrus fruits, which were fixed in a green background frame during the test. The improved YOLOv7 algorithm and the 3D spatial localization algorithm that comes with the depth camera were used on a computer running Ubuntu 16.04 under the ROS (robot operating system) environment. The depth camera identifies and localizes citrus fruits in real time by setting the center point of the binocular camera as the origin of the world coordinate system, the horizontal direction of the origin to the left camera as the X-axis direction, the horizontal direction of the origin to the right camera as the Z-axis direction, and the vertical upward direction of the origin as the Y-axis direction. The distance measurement of the identified citrus fruit was carried out using a tape measure, and the relative spatial position of the target citrus fruit to the ZED binocular depth camera was measured as (Xa, Ya, Za), and compared with the position information returned by the binocular depth camera; a total of 80 measurements were made. The deviation of the spatial position of the target citrus fruit in the Xa measured by the binocular depth camera compared to the manual measurement was 1.02%; the deviation of the spatial position of the target citrus fruit in the Ya measured by the binocular depth camera compared to the manual measurement was 1.57%; the deviation of the spatial position of the target citrus fruit in the Za measured by the binocular depth camera compared to the manual measurement was 2.02%. In addition, the same was also performed for the recognition speed of citrus fruitsAs shown in Figure 12, which was about 4.25 f/s.
Figure 12. Visual recognition testbed for citrus-picking robots.

3.3. Citrus Fruit-Picking Performance Test

From 30 November to 13 December 2023, a field verification experiment of a citrus-picking robot was carried out in the orchard of the Hunan Academy of Agricultural Sciences. The test time was from 1 PM to 6 PM, and different groups of tests were carried out on cloudy and sunny days, respectively. The test fruit tree variety was “Dafen 4”. As shown in Figure 13, 2000 naturally growing citrus fruits were selected for 10 groups of picking experiments. Groups 1–5 were tested on cloudy days and groups 6–10 were tested on sunny days.
Figure 13. Field test environment of citrus-picking robot.
As can be seen from Table 3, the positioning success rate of the picking robotic arm for citrus fruits in the natural environment of the field is 87.15%, and the picking success rate is 82.4%. During the picking test, the unsuccessful completion of positioning and picking was mainly due to the location of citrus fruits beyond the end-effector picking range, and the movement parameters of the joints of the robotic arm had errors which affected the movement accuracy of the robotic arm, leading to the failure of picking.
Table 3. Field performance test results of picking robots.

4. Conclusions

A six-axis autonomous citrus-picking robot system was designed to meet the demand of efficient picking in typical standard orchards in China, which enabled the independent picking operation of citrus fruits in a natural environment. An improved YOLO v7 dense citrus detection model was proposed, which not only makes full and effective use of the interaction characteristics of the global dimension, but also makes the network balance the problem of computing speed and model complexity. According to the physical parameters of the citrus fruit and stem, an end-effector suitable for picking citrus fruit was designed, which effectively reduced the damage associated with citrus-picking. Finally, according to the actual distribution of citrus fruits in the natural environment, a citrus fruit-picking task planning model was established to ensure the efficient and orderly picking operation of the end-effector. The detailed conclusions are as follows:
(1)
In this study, based on the complementary nature of deep learning methods in citrus-picking, targeting the problems of leakage, false detection, and low confidence in citrus fruit region detection, we used the Mobile Net network instead of the CSPDarknet network as the backbone network in the YOLOv7 network model, and incorporated the CA mechanism, which dramatically increases the recognition performance of citrus fruits by picking robots. it can be seen that the improved YOLOv7 model proposed in this paper has a detection accuracy P, recall R, reconciliation mean F1, and average precision AP of 95.27%, 93.28%, 92.88%, and 96.55%.
(2)
A localization test of citrus fruits was carried out, and the data statistics of the binocular depth camera compared to the manually measured spatial position nal deviation of target citrus fruits in X-axis was 1.02%; the depth camera compared to the manually measured spatial position deviation of target citrus fruits in the Y-axis was 1.57%; and the depth camera compared to the manually measured spatial position deviation of target citrus fruits in the Z-axis was 2.02%. In addition, the recognition speed of the citrus fruits was likewise counted, and the recognition speed was about 4.25 f/s.
(3)
The robot was tested for picking performance in the field. The picking robot first identifies and locates citrus fruits through binocular cameras, then controls the robotic arm to move to the picking point, and finally controls the end-effector to complete the picking task. The experimental results show that the success rate of the citrus-picking robot arm is 87.15%, and the success rate of picking in the natural field environment is 82.4%, which is better than the success rate of 80% of the market picking robot. To improve the performance of picking robots in the future, it is necessary to integrate cutting-edge sensing technology, deep learning algorithms, precision mechanical control, and utilize intelligent optimization and adaptive learning with the help of the Internet of Things to achieve efficient and accurate automated picking.

Author Contributions

Software, X.X.; Formal analysis, Y.J.; Investigation, Y.W.; Writing—original draft, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by Hunan Intelligent Agricultural Machinery Equipment Innovation Research and Development Project (Z2023260002414).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, D.; Song, H.; He, D. Research advance on vision system of apple picking robot. Trans. Chin. Soc. Agric. Eng. 2017, 33, 59–69. [Google Scholar]
  2. Xiang, J.; Wang, L.; Li, L.; Lai, K.-H.; Cai, W. Classification-design-optimization integrated picking robots: A review. J. Intell. Manuf. 2023, 1–24. [Google Scholar] [CrossRef]
  3. Vasconez, J.P.; Kantor, G.A.; Cheein, F.A.A. Human–robot interaction in agriculture: A survey and current challenges. Biosyst. Eng. 2019, 179, 35–48. [Google Scholar] [CrossRef]
  4. Bai, Y.; Zhang, B.; Xu, N.; Zhou, J.; Shi, J.; Diao, Z. Vision-based navigation and guidance for agricultural autonomous vehicles and robots: A review. Comput. Electron. Agric. 2023, 205, 107–124. [Google Scholar] [CrossRef]
  5. Li, Z.; Yuan, X.; Yang, Z. Design, simulation, and experiment for the end effector of a spherical fruit picking robot. Int. J. Adv. Robot. Syst. 2023, 20, 17298806231213442. [Google Scholar] [CrossRef]
  6. Wu, H.; Wang, Y.; Shi, Y.; Zhu, Q.; Tan, H.; Wu, Z. Bottom-Up Clustering and Merging Strategy for Irregular Curvature Aero-blade Surface Extraction. IEEE Trans. Instrum. Meas. 2023, 72, 1–9. [Google Scholar] [CrossRef]
  7. Ji, W.; Zhang, T.; Xu, B.; He, G. Apple recognition and picking sequence planning for harvesting robot in the complex environment. J. Agric. Eng. 2023, 55, 1549. [Google Scholar] [CrossRef]
  8. Xu, R.; Li, C. A modular agricultural robotic system (MARS) for precision farming: Concept and Implementation. J. Field Robot. 2022, 39, 387–409. [Google Scholar] [CrossRef]
  9. Wu, H.; Wang, Y.; Vela, P.A.; Zhang, H.; Yuan, X.; Zhou, X. Geometric Inlier Selection for Robust Rigid Registration with Application to Blade Surfaces. IEEE Trans. Ind. Electron. 2022, 69, 9206–9215. [Google Scholar] [CrossRef]
  10. Gao, G.; Guo, H.; Zhou, W.; Luo, D.; Zhang, J. Design of a control system for a safflower picking robot and research on multisensor fusion positioning. Eng. Agrícola 2023, 43, e20210238. [Google Scholar] [CrossRef]
  11. Zhang, X.; Yao, M.; Cheng, Q.; Liang, G.; Fan, F. A novel hand-eye calibration method of picking robot based on TOF camera. Front. Plant Sci. 2023, 13, 1099033. [Google Scholar] [CrossRef] [PubMed]
  12. Bachche, S. Deliberation on design strategies of automatic harvesting systems: A survey. Robotics 2015, 4, 194–222. [Google Scholar] [CrossRef]
  13. Yaguchi, Y.; Takeuchi, K.; Waragai, T.; Tateno, T. Durability evaluation of an additive manufactured biodegradable composite with continuous natural fiber in various conditions reproducing usage environment. Int. J. Autom. Technol. 2020, 14, 959–965. [Google Scholar] [CrossRef]
  14. Chen, X.; Chaudhary, K.; Tanaka, Y.; Nagahama, K.; Yaguchi, H.; Okada, K.; Inaba, M. Reasoning-based vision recognition for agricultural humanoid robot toward tomato harvesting. In Proceedings of the 2015 IEEE/RSJ international conference on intelligent robots and systems (IROS), Hamburg, Germany, 28 September–2 October 2015. [Google Scholar]
  15. Ma, W.; Yang, Z.; Qi, X.; Xu, Y.; Liu, D.; Tan, H.; Li, Y.; Yang, X. Study on the Fragrant Pear-Picking Sequences Based on the Multiple Weighting Method. Agriculture 2023, 13, 1923. [Google Scholar] [CrossRef]
  16. Liu, J.; Peng, Y.; Faheem, M. Experimental and theoretical analysis of fruit plucking patterns for robotic tomato harvesting. Comput. Electron. Agric. 2020, 173, 105330. [Google Scholar] [CrossRef]
  17. Liu, J.; Yuan, Y.; Gao, Y.; Tang, S.; Li, Z. Virtual model of grip-and-cut picking for simulation of vibration and falling of grape clusters. Trans. ASABE 2019, 62, 603–614. [Google Scholar] [CrossRef]
  18. Ji, W.; Zhao, D.; Cheng, F.; Xu, B.; Zhang, Y.; Wang, J. Automatic recognition vision system guided for apple harvesting robot. Comput. Electr. Eng. 2012, 38, 1186–1195. [Google Scholar] [CrossRef]
  19. Cao, S.; Zhao, D.; Liu, X.; Sun, Y. Real-time robust detector for underwater live crabs based on deep learning. Comput. Electron. Agric. 2020, 172, 105339. [Google Scholar] [CrossRef]
  20. Li, G.; Ji, B.Y.; Gu, B.X. Recognition and location of oscillating fruit based on monocular vision and ultrasonic testing. Nongye Jixie Xuebao/Trans. Chin. Soc. Agric. Mach. 2015, 46, 1–8. [Google Scholar]
  21. Mai, X.; Zhang, H.; Jia, X.; Meng, M.Q.-H. Faster R-CNN with classifier fusion for automatic detection of small fruits. IEEE Trans. Autom. Sci. Eng. 2020, 17, 1555–1569. [Google Scholar] [CrossRef]
  22. Ding, X.; Chen, T.; Wei, Y.; He, Z.; Cui, Y.; Yang, Q.C. Design and Implementation of the Positioning and Directing Precision Seeder for Cucurbita Ficifolia Seeds. Appl. Eng. Agric. 2023, 40, 1–14. [Google Scholar] [CrossRef]
  23. Wang, M.; Ning, P.; Su, K.; Yoshinori, G.; Wang, W.; Cui, Y.; Cui, G. Slip-draft embedded control system by adaptively adjusting the battery position for electric tractors. Int. J. Agric. Biol. Eng. 2023, 16, 155–164. [Google Scholar] [CrossRef]
  24. Ling, X.; Zhao, Y.; Gong, L.; Liu, C.; Wang, T. Dual-arm cooperation and implementing for robotic harvesting tomato using binocular vision. Robot. Auton. Syst. 2019, 114, 134–143. [Google Scholar] [CrossRef]
  25. Zhao, Y.; Gong, L.; Huang, Y.; Liu, C. A review of key techniques of vision-based control for harvesting robot. Comput. Electron. Agric. 2016, 127, 311–323. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Zhang, K.; Yang, L.; Zhang, D.; Cui, T.; Yu, Y.; Liu, H. Design and simulation experiment of ridge planting strawberry picking manipulator. Comput. Electron. Agric. 2023, 208, 107690. [Google Scholar] [CrossRef]
  27. Wang, Y.; He, Z.; Cao, D.; Ma, L.; Li, K.; Jia, L.; Cui, Y. Coverage path planning for kiwifruit picking robots based on deep reinforcement learning. Comput. Electron. Agric. 2023, 205, 107593. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.