Next Article in Journal
An Adaptive Assistance Controller to Optimize the Exoskeleton Contribution in Rehabilitation
Next Article in Special Issue
A Case Study on Improving the Software Dependability of a ROS Path Planner for Steep Slope Vineyards
Previous Article in Journal
Variational AutoEncoder to Identify Anomalous Data in Robots
Previous Article in Special Issue
Prototype Development of Small Mobile Robots for Mallard Navigation in Paddy Fields: Toward Realizing Remote Farming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Study on the Feasibility of Robotic Harvesting for Chile Pepper

by
Muhammad Umar Masood
and
Mahdi Haghshenas-Jaryani
*
Department of Mechanical and Aerospace Engineering, New Mexico State University, Las Cruses, NM 88003, USA
*
Author to whom correspondence should be addressed.
Robotics 2021, 10(3), 94; https://doi.org/10.3390/robotics10030094
Submission received: 21 May 2021 / Revised: 15 July 2021 / Accepted: 20 July 2021 / Published: 22 July 2021
(This article belongs to the Special Issue Advances in Agriculture and Forest Robotics)

Abstract

:
This paper presents a study on the robotic harvesting of New Mexico type chile pepper, in a laboratory setting, using a five degrees of freedom (DoF) serial manipulator. The end-effector of the manipulator, a scissor-type cutting mechanism, was devised and experimentally tested in a lab setup which cuts the chile stem to detach the fruit. Through a MATLAB™-based program, the location of the chile pepper is estimated in the robot’s reference frame, using Intel RealSense Depth Camera. The accuracy of the 3D location estimation system matches the maximum accuracy claimed by the manufacturer of the hardware, with a maximum error to be in Y-axis, which is 5.7 mm. The forward and inverse kinematics are developed, and the reachable and dexterous workspaces of the robot are studied. An application-based path planning algorithm is developed to minimize the travel for a specified harvesting task. The robotic harvesting system was able to cut the chile pepper from the plant based on 3D location estimated by MATLAB™ program. On the basis of harvesting operation, on 77 chile peppers, the following harvesting indicators were achieved: localization success rate of 37.7%, detachment success rate of 65.5%, harvest success rate of 24.7%, damage rate of 6.9%, and cycle time of 7 s.

1. Introduction

Chile pepper, as the signature crop of the state of New Mexico (NM), plays an important role for many small acreage growers in the predominately Hispanic and Native American population areas in the state [1,2]. The green chile (Capsicum annuum) crop is entirely hand harvested in New Mexico, requiring a large number of farm laborers during a relatively narrow harvest window. High cost and limited labor availability, as reported throughout the U.S. agriculture sector [3,4], are main contributing reasons for reduced chile pepper production in the state of New Mexico over the past two decades. In 2020, the harvested area for the chile pepper was reported to be 3440 ha (8500 acres) [5], which was significantly declined from 13,962 ha (34,500 acres) in 1992 [6]. Harvest labor costs up to 50 percent of the total production cost in USA. The studies have shown that the cost can be reduced up to 10 percent with mechanization or use of robotic harvesting [6]. Despite the extensive attempts to use mechanized chile pepper harvester as early as 1972 [7], these systems have not been adopted by industry for chile pepper harvest. Mechanized harvesting for the chile, specifically, and other crops, in general, has four major tasks [8]. The first task is dividing that the machine separates the fruit from the bushes and leaves as it goes from one plant to another, to avoid damage to the plant and the harvested fruit. Second, being the most important is the detachment of the fruit from the plant. The third step is to grasp and convey the fruit, after being removed from the plant. Fourth is the transportation from the field to the post-processing site and the consumer market. Detachment task in robotic harvesting has extra importance because it is challenging, crop-specific, and has the highest effects on the economic value of the fruit.
In New Mexico, NM-type red chile and paprika, NM-type long green chile, and Cayenne peppers are the three main types of chile peppers that contribute the most to the overall chile industry. Each of these three types of chile requires a different type of mechanization for harvesting, as they differ in size, shape, and uses. Additionally, the product damage for the chile peppers which are sold as fresh market item, like green chile pepper, is not acceptable, which makes the mechanization of harvesting more challenging. Heavy-duty and large harvesting machines can perform in larger field areas where only one harvesting cycle is needed for the crop, as the mechanized harvesters, mostly destroy the plant in their first harvesting process [9]. Moreover, mechanized harvesting is not selective harvesting, so it brings up the bushes and sometimes soil with the fruit. On the other hand, cultivation in a greenhouse and on small farms has been increasing with time, where these harvesters cannot be used because of their larger size. These facts point to the necessity of using other advanced technologies, such as robotic harvesters, for selective and adaptive harvesting, which should be small enough to work in a canopy and harvest from a variety of plants in greenhouses. Such robotic harvesting solutions can also be used in fields.
Robotic systems have been used in farming systems to address growing global challenges in agriculture [3,4,10]. Harvesting has been significantly affected by a consistently shrinking and increasingly more expensive labor force; therefore, there has been substantial research related to the design and development of field robots for harvesting [10,11]. Robotic arms (manipulators) are designed or selected from commercially available platforms with degrees-of-freedom (DoF) varying between two to seven with the majority of them having three [10,12,13,14]. However, none of the past research in harvesting robots has offered systematic analysis and design for the selection of the number of DoF [10]. On the other hand, robotic end-effectors are mostly custom made and specialized for specific crop harvesting. Some of these robotic end-effector grasp and hold the fruit using suction cups [15,16] or a combination of suction cups and robotic fingers [17,18]. The detachment process is carried out using thermal cutting, scissor cutting, or physical twist and pulling for detaching the fruits at the stem [12,14,19].
There have been a lot of research efforts in robotic harvesting for fruits, which are hanging on small plants near the ground [20], including berry and tomato [14,21,22,23]. The fruit removal mechanism adopted by these robots includes: (1) cutting the stem [13,15] and (2) gripping or grasping the fruit and applying a normal force to destem. For example, Gunderman et al. [21] used a soft gripper for berry harvesting which takes into account the fingertip force, fruit size and shape, and nature of the plant. Hohimer et al. [24] recently presented a 5-DoF apple harvesting robotic arm with a soft gripper as end-effector and the detachment mechanism. Some of the other robotic systems use the manipulator’s movement to apply force for the detachment [16,25], rather than just using the torque by the wrist joints. Despite the variation in their core concept, all of the designs considered the requirement of delicate grasping to avoid fruit damage [26,27].
All these harvesting techniques, which grip the fruit and pluck it by applying force have been mainly applied for trees and rigid stem plants. However, for delicate plants, this method can create a lot of damage to the unharvested fruit and plant itself. Harvesting the fruit by cutting the pedicle (the stem holding the fruit with plant) is very useful when the subject plant is very delicate and cannot sustain a lot of force from the robotic manipulator. In this method, there is no extra force applied on the plant, rather the robot simply cuts the pedicle and takes off the fruit. For instance, Liu et al. [15] developed a litchi picking robot with a cutting mechanism. The pneumatic force is transmitted to the cutting blades through a tendon and rigid bar. Two designs are discussed and compared, where a detailed quality analysis comparison is shown for harvesting, through hand picking, manual mechanized picking, and the harvester picking. Laser cutting is also studied as the cutting technique for detachment of spherical fruit by Liu et al. [28]. Jia et al. [29] developed an integrated end-effector which can grip the fruit and then cut the stem, as a single device, using a clutching mechanism. The actions of holding and cutting are designed as a logical sequence action.
Task and motion planning are also essential parts of robotic harvesting systems. A few works studied task planning strategies for harvesting which are mainly based on coverage path planning to pick all available fruits in a scene [16] or to minimize the required time for moving between fruits [30]. The direct displacement towards the desired position of the end-effector is the most common approach for path planning of the robotic manipulators. This approach was achieved using position-based control [19] and visual feedback control [31]. Other control algorithms, like fuzzy control [19], combined PD and linear quadratic technique [32], and impedance control method [33], for coordinating robotic manipulators and grasp modulation, have been employed.
Despite all of these efforts for harvesting different crops [10], there has not been any effort towards design and development of robots/intelligent machines for harvesting chile pepper. Aforementioned issue in regards to mechanized machines has been a huge factor in preventing transition to mechanical harvest, since human laborers can selectively pick marketable green fruit, while mechanical harvesters completely strip all fruit, including immature pods that could provide for a later harvest [8,9]. Robotic harvesters could potentially selectively harvest marketable green chile fruit and successful implementation of robots for NM green chile could reverse the decline of acreage in the state (from 38,000 acres in mid-1990s to 8400 in 2018). However, the complexity of the chile pepper crops environment would be a challenging factor regarding hardware design, computer vision, and motion/control algorithms to achieve an effective robotic harvester. The best strategy is to break down the complexity of the problem.
This work presents a feasibility study of using a robotic arm and customized end-effector for harvesting chile pepper. The structure of green chile pepper plant and random distribution of fruits all around it would require use of a more dexterous robotic arm with 5 to 7 degrees-of-freedom. A rigid link commercial robot Tinkerkit Braccio Arm with 5 DoF is modified and used along with a cutting mechanism to detach the chile pods from the plant in a laboratory setting. Similar to the other low-hanging crops on small plants, like berries and tomato, the cutting mechanism was considered as the primary option for the end-effector. The scope of this paper is to cut the fruit stem with a cutting mechanism while to keep the plant and other unharvested fruit safe. A MATLAB™-based software is developed which localizes the fruit in 3D space, based on identification of the fruit in the image by the human operator. For this purpose, Intel RealSense Depth Camera D435i is used for collection of stereo and RGB images. The reachable and dexterous workspace of the robot was studied with respect to the spatial distribution of chile peppers to determine the optimal configuration of the robot for maximum harvesting rate. Motion planning in Cartesian space was carried out for manipulation and harvesting tasks. The operation of the robotic harvesting for chile pepper was examined in a series of bench-top testing for a total of 77 chile peppers. The obtained data were analyzed, and harvesting metrics were determined, to assess the proof-of-concept and feasibility of using the robotic system for chile pepper harvesting.
The organization of this paper is as follows. Section 2 gives an overview about the overall robotic harvesting system and its components. In Section 3, the details of the Harvesting robotic arm, forward and inverse kinematic, workspace analysis, customized end-effector design, and prototyping will be discussed. Section 4 introduces the vision-based fruit localization, camera setup, definition of the frames, and the error analysis. Section 5 presents the study of chile pepper’s geometrical features and distribution of chile pods on the plant grow in a pot. In Section 6, motion planing strategy and the harvesting scheme will be introduced. Finally, Section 7 discusses the results of the experimental harvesting and obtained harvesting indicators.

2. Robotic Harvesting System

Harvesting robot consists of a 5-DoF robot arm, a cutting mechanism, an Arduino Due board, an Intel RealSense Depth Camera D435i and a MATLAB™ running machine. The schematic of the harvesting system and physical lab setup is shown in Figure 1 and Figure 2, respectively. The Intel RealSense Depth Camera D435i takes both the Red-Green-Blue (RGB) and depth images and transfers it to our MATLAB™ code. The whole bench top, consisting of the test chile pepper plant for harvesting and the harvesting robot, is in the viewing frame of the camera. The MATLAB™-based graphical user interface shows only the RGB image to the user for identification of the ripe chile pepper. The human operator identifies and selects one chile pepper in the image, as shown in Figure 3, by a mouse click on that point in the image. On the back end, MATLAB™ registers both the RGB and the depth images together and generates a mapping between them. Based on user identifying the chile pepper in the image, the MATLAB™ program takes the depth and point cloud data for that specific pixel, which is selected with the mouse click, using Intel RealSense Software Development Kit (SDK) functions. Then, it transforms these data into the robot coordinate frame using designated homogeneous transformations.
Although the human operator will not exactly choose the best place to be cut, the operator does provide an approximate location for the robot to know where is the connecting point of the fruit to the stem as a reference point. Based on the 3D localization obtained from the image, the robot will move its end-effector to a point (offset from the human operator’s selected point) on the stem determined according to a pre-set offsets in x-y-z direction. These offsets were determined based on the prior analyses of the fruit geometrical features (e.g., average stem length) presented in Section 5. Then, the localization information will be communicated to the Arduino Due board installed on the harvesting robot. This localization information consist of three coordinates data. Given the current joint space position and Cartesian space position of the robot, the arduino program assesses whether this location is within the reachable space of the robot. If the point is in the reachable space, the robot approaches the identified fruit from the outside of the plant, to avoid or minimize collision of the end-effector with any branch or other obstacles within the plant. The cutting mechanism cuts the stem which is connecting the fruit with the plant. The whole operation is considered as a complete harvesting cycle.

3. Harvesting Robot

A systematic design starts from the study of the application requirements, limitations, and valid assumptions. The following subsections give an insight on the performance and limitations of the robot arm being used for this study, with respect to the challenges for chile pepper harvesting.

3.1. Braccio Robotic Arm

Braccio, shown in Figure 4, is a five degrees of freedom (5 DoF), low-cost robot arm used for our preliminary study on the robotic harvesting of chile peppers and proof-of-concept. It is based on DC Servo Motors which are controlled by their internal circuitry in a discrete manner based on the input reference signal send by a micro-controller. In this work, Arduino Due Development Board is used as microcontroller for the hardware. Starting from the base to the end-effector, the joints of the robot arms are named as base, shoulder, elbow, wrist pitch and wrist roll, respectively. All the motors at the joints have a range of motion from 0° to 180° degrees except the shoulder joint (2nd joint), which can only move from 15° to 165°. The limitation of this joint is because of the mechanical design and structural limitations, not the motor’s extreme positions. The maximum reach of the robot arm is 400 mm, while the maximum payload capacity is 150 g for an operating distance of 320 mm and 400 g for minimal operating distance.

3.2. Forward Kinematics

Forward kinematics of the robot is derived in the form of linear homogeneous transformation using Denavit–Hartenberg (D-H) parameters. The joint space is defined so that the zero position of each joint is in the center of its space. It is important to note that the Braccio comes with 180 ° servo motors, which give an actuator space of 0 to 180 ° . The joint space selected, following most common notation, is 90 ° to 90 ° . The frames assignment is shown in Figure 5. The D-H parameters are written based on the shown frames and tabulated in Table 1. A homogeneous transformation can be written as
T E 0 = s 1 s 5 c 1 c 5 s 234 c 5 s 1 + c 234 c 1 s 5 c 1 s 234 c 1 ( a 3 s 23 + a 2 s 2 + d 6 s 234 ) c 1 s 5 c 5 s 1 s 234 c 5 c 1 c 234 s 1 s 5 s 1 s 234 s 1 ( a 3 s 23 + a 2 s 2 + d 6 s 234 ) c 5 s 234 s 5 s 234 c 234 a 3 c 23 d 1 + a 2 c 2 + d 6 c 234 0 0 0 1 ,
where c n = c o s ( θ n ) , s n = s i n ( θ n ) , c 234 = c o s ( θ 2 + θ 3 + θ 4 ) , s 23 4 = s i n ( θ 2 + θ 3 θ 4 ) , s 23 = s i n ( θ 2 + θ 3 ) , and c 23 = c o s ( θ 2 + θ 3 ) . a 2 , a 3 , d 1 , and d 6 = d E are defined in the D-H parameters (see Table 1).
Figure 5. Braccio schematic, frames, and joints definition.
Figure 5. Braccio schematic, frames, and joints definition.
Robotics 10 00094 g005
Table 1. D-H Parameters of Braccio robot arm.
Table 1. D-H Parameters of Braccio robot arm.
i α i 1 a i 1 d i θ i
11800−70 θ 1
29000 θ 2 90
301250 θ 3
401250 θ 4 90
59000 θ 5
E1800−1500
All lengths in mm, while angles are in degrees.
This forward kinematics is used for MATLAB™ simulations and implementation in Arduino.

3.3. Inverse Kinematics

The inverse kinematics of this 5 DoF robot can be written in the geometric form, where the inputs are three coordinate values for position, called as x, y, and z, and two orientation angles called as pitch (denoted as β ) and roll (denoted as γ ). All the joint angles are denoted as θ 1 , θ 2 , θ 3 , θ 4 , and θ 5 . The coordinate frames are shown in the Figure 5. The base angle θ 1 can be determined as
θ 1 = tan 1 ( y x ) .
Take r as the distance from the end-effector position in x-y plane (projection on this plane) ( x , y ) to the origin of the base frame of the robot, r = x 2 + y 2 . Defining a new plane comprising of r in x-y plane, and z- a x i s , denoted as r-z plane, the solution for next three joints can be solved as a 3-DoF planar robot.
For a specified pitch angle β , the location of the 4th joint in r-z plane ( r , h ) can be found as r = r d E c o s ( β ) and h = z d 1 d E s i n ( β ) . Now, all the sides of the triangle formed by joint 2, joint 3, and joint 4 are known, so the angles ρ , ψ , and ϕ (shown in Figure 6) can be determined by the rule of cosines and the joint angles can be written as
θ 2 = ρ + ψ θ 3 = π ϕ θ 4 = β θ 2 θ 3 θ 5 = γ .
The last angle θ 5 is independent of position and pitch β .

3.4. Reachability and Dexterity

As mentioned earlier, the robot has five degrees of freedom, which limits its dexterity in the yaw orientation, being the rotation about z- a x i s , and referred to as α . For any position in reachable space, no value other than zero is possible for yaw, when taken in end-effector’s moving frame. Roll, being the rotation about x- a x i s , and referred to as γ , is independent of the end-effector’s position so, any roll orientation is practically possible. The pitch, rotation about y- a x i s , denoted as β , affects the dexterous space considerably enough, to be studied. It is mostly because of the limitations of the joint movements.
A MATLAB™ program is developed to plot and visualize the reachable workspace of the robot. In this program, only the front half of the robot is plotted, to make the plot more elaborate. The actual workspace is twice (same as the other side of the robot) than what is shown here. The dexterous space of the robot is mainly dependent upon on the pitch angle. For a specified pitch angle, there would be a different dexterous space. Considering these limitations from the hardware, it is needed to define the minimal requirements for the application. Although the ideal pitch angle for the harvesting purpose using Braccio is 0 ° , but a range of ± 25 ° was found practically good enough to harvest a chile pepper from the plant. It is found that the dexterous space for β = 0 ° is incredibly small, but, with the flexible range of β = ± 25 ° , the dexterous space is adequate to capture most of the chile peppers from the plant. This will be discussed in detail in the preceding sections. Figure 7 shows the reachable and dexterous space of the Braccio robot arm in the form of the point cloud data. For a better understanding, the planar view is also shown in Figure 7b, where the blue points show the space where the robot is reachable with β = 0 ° , and the green points show the space which is reachable with any pitch angle β in the range of β = ± 25 ° else β = 0 ° . The space covered by magenta points is the overall reachable workspace with any pitch angle β .

3.5. Repeatability and Accuracy

Repeatability and accuracy are very important parameters to assess the robot’s performance. The repeatability and accuracy test was performed on the robot by giving a target location in 3D space and measuring the actual location of the end-effector where it reaches. It was tested for three different points in the 3D workspace, ( 100 , 100 , 250 ) , ( 100 , 100 , 250 ) , and ( 150 , 0 , 320 ) ; called as point 1, point 2 and point 3, respectively. The three points are selected so that the performance can be assessed in both quadrants, as well as on the x- a x i s , which lies on the middle of the robot’s workspace. A total of 33 samples were recorded, having 11 for each point, to assess the inherent performance parameters. The overall average of the difference found in target and reached point is denoted as μ . The standard deviation for the same point is a metric parameter for repeatability. For each point, standard deviation is denoted as σ 1 , σ 2 , and σ 3 . The Table 2 shows the results.

3.6. Detachment of Chile Fruit

3.6.1. Computer-Aided Design

It was discussed earlier that, for delicate fruit, cutting of pedicle is the most effective technique for detachment from the plant, in terms of minimum damage to the plant. The detachment technique used for harvesting, in this work, is the cutting of stem through sharp edge blades which are actuated with a DC Position Servo motor. This customized cutting mechanism is mounted on Braccio robot arm as the end-effector. The cutter mechanism needs to be designed considering the constraints for size, weight, and mode of operation. These constraints are defined by the capabilities and limitations of Braccio robot arm. The maximum payload capacity of Braccio robot arm is 150 g, so the weight of the designed end-effector was limited to 150 g increment from the original end-effector. Size of the end-effector was also considered so that the minimum collision with the obstacles can be achieved. The torque of the motor held with the gripper of Braccio robot arm is not enough to cut a stem of diameter ranging from 4.7 mm to 6.3 mm. Moreover, the Braccio’s wrist motor is one of the main bottleneck in the limitation of payload capacity. Thus, both motors, the one responsible for roll rotation and the gripper, called M5 and M6, were replaced with metal gear and metal horn servo motors, DS 3218 DC Servo Motor, with the maximum torque of 2 Nm.
As shown in Figure 8a, the cutter is designed such that one blade is fixed with the ground of end-effector and the other blade is moving about a pivot which lies within the body of the blade. This way, the maximum cutting force can be achieved depending on the maximum torque provided by the motor. Additionally, the torque is enhanced by 5.5 times through a set of pinion and gear (with 12 and 54 teeth, respectively) as the transmission method. This helps to drastically increase the cutting success rate. The CAD model of the cutter developed in SolidWorks™ is shown in Figure 8a. The torque at the pivot of the blade come out to be τ c = 5.5 τ m = 11 Nm, where τ c is the torque at cutter, and τ m is the maximum torque of the motor. The maximum torque is applied when the motor is operated at 6.0 V.

3.6.2. Prototyping and Manufacturing

The designed cutter mechanism was fabricated using a stereolithography (SLA) printer Form 3 (Formlabs, Somerville, MA, USA). For the gears, engineering grade Rigid 400 from Formlabs [34] is used. The blades used for this design are commercially available as replacement blades for gardening pruners. The mounting fixture of the mechanism was designed considering that it should be installed on Braccio robot arm as end-effector. The assembled cutting mechanism installed on the robot’s end-effector is shown in Figure 8b.

4. Vision-Based Fruit Localization

A vision-based fruit localization was used to determine the coordinates of the chile pepper and its stem. Location of an object in camera frame { C } can be estimated in 3D space, given the camera intrinsic and extrinsic. For this purpose, Intel RealSense™ Depth Camera D435i is used, which provides an RGB and depth images. Additionally, the camera is equipped with an Inertial Measurement Unit (IMU), which is used to determine the relative orientation of the camera frame { C } with respect to the world frame { W } . Vision-based identification or classification of the chile fruit is not the scope of this research, so RGB image plays an important role for the user (human operator) to identify chile pods in the image and click on its location with possibly minimal error, as compared to a greyscale depth image. The camera provides 3D image, as well, which may be taken as 3D point cloud data in correlation to every depth image pixel. By combining both RGB and depth image, we can co-relate an image pixel with its 3D location in the camera frame. The location of the targeted point is required to be known in the robot frame { R } . Another frame called the world frame { W } is introduced at a known point in the robot frame. All these frames and the camera setup is shown in Figure 9.
The robot is placed on a lab bench top with its coordinates shown in Figure 9, and the coordinate frame is denoted as { R } . The X Y plane of this frame is aligned with the bench top. A global frame, called world frame and denoted as { W } , is introduced. Location of camera and the origin of the robot frame can be measured in { W } as there are fixed for the whole operation. Z-axis of frame { W } is perfectly aligned with gravity whether the frame { R } may not be aligned because of inherent inclination of the table and the floor. The rotation matrix R W C describes the camera frame with respect to the world frame. The accelerometer readings ( a x , a y , and a z ) were used to solve for the roll and pitch angles ( ϕ and θ ), and we assumed the yaw angle ( ψ = 0 ) known due the fixed and known pose of the camera with respect to the benchtop test setup in this study. We externally verified the yaw rotation by aligning the horizontal grid line (see Figure 9).
a = R x ( ϕ ) R y ( θ ) R z ( ψ ) T g ,
where a = [ a x , a y , a z ] T and g = [ 0 , 0 , 1 ] T are normalized accelerometer readings and gravity vectors, respectively. Solving for the roll and pitch angles ( ϕ and θ ) gives:
tan ( ϕ ) = a y a z and tan ( θ ) = a x a y 2 + a z 2 ,
which yields:
sin ( ϕ ) = a y a y 2 + a z 2 cos ( ϕ ) = a z a y 2 + a z 2 sin ( θ ) = a x cos ( θ ) = a y 2 + a z 2 ;
substituting these terms back into R W C = R x ( ϕ ) R y ( θ ) R z ( ψ ) with ψ = 0 yields:
R C W = a y 2 + a z 2 a x a y a y 2 + a z 2 a x a z a y 2 + a z 2 0 a z a y 2 + a z 2 a y a y 2 + a z 2 a x a y a z ,
where
a x 2 + a y 2 + a z 2 = 1 .
In a practical real-world setting with a moving camera, as opposed to the current case, an extra orientation measuring sensor (e.g., electronic compasses or 9-DoF IMU sensor) will be used to calculate the required rotation angles and the rotation matrix.
In Equation (4), a x , a y , and a z are the normalized accelerations with the units of g being the gravitational acceleration. When camera is setup on the tripod, these values can be taken once from the Intel RealSense Viewer software and entered into the MATLAB™ program. The origin point of the world frame was marked on the table and can be seen in the camera image. Using a measuring, the 3D location P C o r g W of camera frame origin can be extracted in the world frame. The homogeneous transformation from { W } frame to { C } frame can be written as T C W = R C W P C o r g W 0 0 0 1 .
The rotation of the robot frame with respect to the world frame takes into account any tilt in the table top. There is no rotation around the z- a x i s because both the frames are on the table top. The table tilt angles θ x and θ y about the x- a x i s and y- a x i s , respectively, were determined once, using an inclinometer, after the table had setup. This rotation matrix can be written as
R R W = R y ( θ y ) R x ( θ x ) ,
where
R y ( θ y ) = c 0 s 0 1 0 s 0 c and R x ( θ x ) = 1 0 0 0 c s 0 s c , where s = s i n θ , c = c o s θ .
The position of the world frame W is already known in the robot frame. Thus, the transformation is given as:
T W R = R R W T P W o r g R 0 0 0 1 .
The combined homogeneous transformation, mapping the camera frame into the robot frame, is obtained as follows: T C R = T W R T C W .
Given the position of any point P C in the camera frame, that can be transformed into the position of the same point in the robot frame. This can be written as:
P R = T C R P C .

4.1. Performance of 3D Localization and Camera

A test setup was developed to measure the accuracy of the 3D location estimation. Figure 9 shows the test setup with a known grid marked on the table top. As the measurements are estimated in 3D, there must be some known position elevated from the table’s surface. A grid marked bar was designed with a flat surface which may be placed at any known point in 2D plane to raise its z coordinate value. Note that the vertical bar should be at right angle with the table, but it may not be aligned with gravity vector, as the inclination of table has already been accommodated in the transformations. Following are the possible sources of the error in the 3D localization system. The overall error in the measurement system is given in Table 3.

4.1.1. Human Operator Error

As mentioned previously, the human operator identifies chile peppers in the image and additionally by clicking on the fruit location it provides a reference point for the robot to position its end-effector for the detachment task. The human operator error, in this case, is due to either the challenge of repeatability of clicking on the same pixel on the image (repeatability) or clicking on a point off the fruit (accuracy). This is very important factor since our robot positioning of stems relies on the reference point location provided by the human operator. Thus, human error is one of the most contributing factor in the integrated error found in this vision-based localization measurement system.

4.1.2. Depth Camera

The depth camera installed does have an inherent error in the depth measurement. This error is unsystematic and needs to be taken as the Original Equipment Manufacturer (OEM) claims. It should be considered that up to 2% error may be there in the measurement of depth when measured in the range of 2 m. Most of these error contributions can be minimized by taking an average of multiple readings. Taking multiple readings is only possible when the image frame or the scene being captured, is stationary. For moving frames, reducing this error can be a challenging task.

4.1.3. Accelerometer

Camera is always in stationary condition for this specific application. In this case, the only acceleration component, that the embedded accelerometer can read is the gravitational vector. The gravitational vector is divided into three Cartesian components, which are normalized and used for orientation of the camera, as given in Equation (4). But there is a known issue with this estimation of orientation. Any rotation about the gravitation vector does not reflect in the orientation matrix. This part was externally verified before taking measurements, by ensuring that each horizontal (along x-axis of the robot frame) grid line in the camera frame should be perfectly horizontal in the frame. This means that all the points on each horizontal line should have same x-pixel value in the image frame. Due to this, there may have some random error in the external verification.

5. Study of Chile Fruit

5.1. Dimensions and Anatomical Features of the Chile Fruit

A data collection and analysis on the size and form factor of the New-Mexico type chile was performed based on 27 chile peppers. The considered parameters are maximum diameter D, projected length L, diameter of the stem d, and the available length of the stem for cutting purpose l. These notations are shown in Figure 10, as well.
The analyzed data in form of histogram are shown in the Figure 11. More than 90% of the chile has a diameter in the range of 24 mm to 38 mm with an average of 29.4 mm and the standard deviation of 5.3 mm. Average length of chile pepper is 113.4 mm with a high variation, showing standard deviation of 21.1. Stem diameter is very important to know for designing a cutting mechanism. The average stem diameter was found to be 5.6 mm with the maximum of 7.5 mm. Having more available length of stem means it is easy to reach the chile even with positional errors caused by either mechanical hardware or software. The results show the average available stem length of 22.1 mm with a significant variation of 12.2 mm. However, in some of the anomalous cases, the fruit may not have any available stem. The analysis results are summarized in Table 4.

5.2. Distribution of Chile Fruit on the Plant

The spatial distribution of the chile pepper fruit on the plant was studied as a crucial factor with respect to the robot’s reachable and dexterous space. If the robotic arm was mounted on a mobile robot, then this problem would be seen from some different perspective, but, in the current scenario, it will be discussed based on the fixed location of the robot arm. Considering the bottom center of the chile pot to be ( 0 , 0 , 0 ) , locations of 57 chile peppers were recorded using the 3D location estimation system used for fruit localization in Section 4. The distribution of the chile fruit with respect to the zero location is shown in Figure 12b. Through a MATLAB™-based program, an ideal location for robot is also calculated so that the maximum chile fruit is in the dexterous space (for β = ± 25 ° ) of the robot. This optimum location of the robot was found out to be ( 180 , 0 , 120 ) with respect to the world frame. Figure 12c shows all the chile fruits to be enveloped in the dexterous space of the robot when the robot is located at the optimum location.

6. Motion Planning and Harvesting Scheme

This section describes the flow of algorithm followed for harvesting, as shown in Figure 13, along with the motion of the robot in the harvesting operation.

6.1. Motion Planning

Trajectory planning was carried out based on a straight line path in Cartesian space where the pitch orientation of wrist was specified to achieve the desired position and orientation for harvesting chile pepper pods. To generate, verify, and simulate the robot’s motion, a MATLAB™-based simulation program is developed, so that the proposed scheme can always be tested in the simulation. Based on Equation (1), and the inverse kinematics, some commonly used functions, like r e a c h O b j e c t , are developed. These functions mainly solve the joint angles of the robot for a specific object’s pose in 3D space. Current joint angles and the end-effector’s position and orientation are always kept in the program, so that robot can make a linear trajectory from the current point to the target point in Cartesian space. The preference is to have one kind of inverse kinematics solution for all the intermediate points; therefore, motions in the joint space with drastic changes can be avoided. Figure 14 shows an example of path planning and operation of the Braccio robotic arm in the MATLAB™ simulation for the target point ( 250 , 250 , 100 ) and the initial position ( 47 , 0 , 233 ) . The corresponding joint angles are illustrated in Figure 14b, which verifies the smooth motion of the robotic arm in the joint space, with the joint angles in step time. The resulted path is provided to the Arduino algorithm for operating the actual robotic arm.

6.2. Harvesting Program

The two-part program runs on MATLAB™ and Arduino simultaneously, as the schematic of the program shows in Figure 13. The MATLAB™ program takes a set of RGB and depth image frame and registers them through the SDK functions. Based on the identification and click on an image pixel, the pixel information is retrieved from the image data and its corresponding depth and 3D point data, from the 3D point cloud built by the SDK function. This 3D location of the chile fruit is in the camera frame. Using Equation (7), the location is transformed into the robot frame. Using serial communication between the MATLAB™ program and Arduino Due board, the location of the chile is communicated to the Arduino, which instructs the robot to reach the specified location and cut the targeted chile.
Arduino program is responsible for the solution of joint angles for the given 3D location. Using inverse kinematics of the robot, this program finds the optimal joint space position of the robot, following a linear trajectory, in the same way as developed in MATLAB™ simulations. Due to the limitations of the Braccio discussed earlier, the dexterous space for one specific pitch angle, β , is very small. In the application of chile harvesting, the ideal pitch angle is β = 0 ° , but it is acceptable for some range called as tolerance in pitch angle. It has been observed, with experimental studies, that the max tolerance can be up to ± 25 ° . An iteration-based algorithm is shown in Figure 13, which solves the solution for the best pitch angle, starting from the ideally required pitch angle. The described harvesting schemes and the operation of the robot was examined in a series of bench top laboratory testing.

7. Results and Discussion

A study has been conducted through a series of experiments to verify the performance, show feasibility of using robotic system, and formulate the potential factors for unsuccessful cases. Generally, performing this study in an indoor (Laboratory) setting has eliminated some of the common challenges in outdoor settings due to the controlled environment, such as working in a structured environment (fixed table and fixed camera view), consistent and controlled light, and disturbances due to wind. Some of the other simplifying assumptions we made in this work include: (1) fruits are fully exposed to the computer vision and the robot (we moderately trimmed the leaves); (2) average geometrical features of chile pepper pods were used for the robot positioning; and (3) the stems are vertically oriented.
The cutting mechanism and technique is proper for the hanging fruits only when the pedicle is separated from other stems and plant objects, and the length of the pedicle is long enough to selectively cut it. As shown in Figure 15a, in some cases, the length of the pedicle is short for the cutting mechanism to selectively cut it for detachment. A higher precision robot may be able to do this job, provided that the width of the cutter is also small enough. Another issue is that, in some cases, there is a big cluster of fruits on the same point in the plant (Figure 15b). First, it is very hard for the vision-based localization system to capture the exact chile pepper’s 3D location; secondly, it is a hard job for the cutting mechanism, as well, to cut the correct chile, in that case. It may increase the damage rate, as well. Another challenging case is shown in Figure 15c, where the chile pepper pedicle (target stem) is very close to the main or sub-branch of the plant. In this case, the robot may damage the plant and, eventually, all the unharvested fruit on that sub branch.
In addition to the aforementioned limitations, the 5-DoF robot can only reach to chile peppers which is approachable from the center of the robot without any obstacle in pathway. The main stem of the plant is one major obstacle, in such a case, when the robotic manipulator is static on a platform. All the chile pepper which have the main stem of the plant in the pathway of robot, cannot be localized by the robot. Using, the same data of 57 chile pepper, as in Section 5, a MATLAB™ simulation is run to see the effect of the main stem of the plant in the pathway to the chile pepper. Figure 16 shows that, out of 57 chile peppers, 21 (which makes it 36.8%) are in the region which not reachable for the robot because of the stem. The simulation shows that the orientation of the plant may change the percentage value. But, in this study, the orientation of the plant was not intentionally selected to increase the success rate; rather, it was arbitrarily chosen.
Chile harvesting was performed by an image click on the MATLAB™-based program, which finds the location of the chile and sends the coordinates to Arduino. The same test setup shown in Figure 2 was used. The robot was raised 120 mm high reference to the chile pot, through a platform. The chile pots were placed at 180 mm at x-axis in the robot frame. These values were found out as the best optimum location of the robot where it can reach maximum number of chile pepper as per the chile distribution data. Because the camera is fixed at one location, some of the view-blocking leaves were trimmed, in order to see all the chile pepper in the image frame. An example of harvesting operation is shown in Figure 17. For the recorded videos of the operation, please refer to the Supplementary Materials, Video S1.
A number of trials were performed to asses the localization success rate, detachment success rate, harvesting success rate, cycle time, and damage success rate. These performance metric are defined by Reference [10] for harvesting in a canopy. They are interpreted for lab testing as follows:
  • Fruit localization success [%]: The number of successfully localized ripe fruit per total number of ripe fruit taken for testing in the lab setup.
  • Detachment success [%]: The number of successfully harvested ripe fruit per total number of localized fruit.
  • Harvest success [%]: The number of successfully harvested ripe fruit per total number of ripe fruit taken for testing in the lab setup.
  • Cycle time [s]: Average time of a full harvest operation, including human identification, localization, and fruit detachment.
  • Damage rate [%]: The number of damaged fruit of pedicles per total number of localized ripe fruit, caused by the robot.
Localization is considered correct only if the stem is found to be in the center of two blades of the cutter. Based on 77 ripe fruit taken for harvesting, the summary of harvest testing results are shown in Table 5. For more details and the collected raw data used for these calculation please refer to the Supplementary Materials, Tables S1 and S2.
Comparison of performance of a robotic harvester to human laborer is a valid point which can indicate whether the use of robot is effective/efficient and economical or not. However, to the best of our knowledge, direct information regarding hand-pick harvesting performance indicators has not been collected. Although, based on our personal experience in the lab, the human clipping, for example, is faster but taking into account the burden of this repetitive task and the capability of robots to perform repetitive tasks in a longer period of time compared to human laborer, the obtained 7 s cycle is justified. Additionally, the average cycle time for different harvesting robots, reported in Reference [10], is 33 s with a large range of 1 to 277 s, which makes our robotic harvester’s cycle time performance to be in a reasonable range.
Based on the previous studies, in this matter, 6.9% is an acceptable rate for fruit damages using mechanized/robotic harvesting. Bac et al. [10] reported on average the fruit damage rate for robotic harvesters (for different crops) is 5% with the low of 0% and high of 13%. Particularly, the fruit damage rate was reported for chile pepper mechanized harvester as low as 7% to 11% which is stated as “acceptably low” [9]. In some other aggressive mechanized harvesters, the fruit damage rate was reported as high as 20% [8]. In our future work, we will use in-hand visual servoing and related motion planning to improve this harvesting factor.
In this study, the localization of stem was carried out by a combination of the human operator fruit detection and robot pre-motion planning algorithm. Therefore, the real localization which must be vision-based had not be implemented. In this study, we defined the localization in a slightly different way, and it was considered correct only if the stem is found to be in the center of two blades of the cutter. As it is mentioned, this is the result of the fruit identification by the human operator and other pre-adjustment in the robot motion planning based on our knowledge from the fruit geometrical analysis section. Despite the simplified assumptions, this task was facing a higher rate of failure compared to the other factors due to more complexity of stem and plant structure; please refer to Figure 15. One solution to resolve this issue and improve the rate is to use an in-hand camera (camera mounted at the wrist) to provide real-time and close-up information about the stem geometrical feature, and then the robot would be able to perform on-the-fly adjustment to position the cutter in an optimal position for a successful detachment.
The obtained results for robotic harvesting using a 5-DoF robot show promising for future studies based on the presented feasibility study and the observed challenges, which are specific to the case of chile pepper harvesting. The results indicate that using a robotic arm with more degrees-of-freedom could improve the chile pepper harvesting performance. Moreover, the outcomes of this study indicate the importance of mobility of manipulator to achieve higher harvesting rate. For example, if the same manipulator is mounted on top of a mobile robot, then it can reach all the plants from all sides, which would increase the localization and harvesting success rate. To achieve fully autonomous harvesting robot for chile pepper, our overarching goal, integration of two major components into the current system are needed. First is computer vision which requires fruit identification and localization. As many robotic harvesting applications rely on color differences, it will be critical to establish features capable of distinguishing between the green foliage of the chile plants and the green chile pods themselves, as in Figure 3. Work in detection of other crops with a similar identification challenge, such as green bell peppers [35,36] and cucumber [37], provides some guidance on identifying green fruits based on other features (e.g., slight variations in color or texture (i.e., smoothness) or spectral differences in NIR). Leveraging machine learning and the capabilities of convolutional neural networks (CNNs), which have shown excellent performance in image classification and segmentation, could overcome this challenge. Second, from the standpoint of robot operation, using more advance motion planning and control algorithms are required to perform them.

8. Conclusions

This paper presents a feasibility study on the robotic harvesting for chile pepper, in a laboratory setting, using a 5-DoF serial manipulator. A simple and easy to develop 3D location estimation system was developed with considerably minimal mean error which can be efficiently used for various research applications which involve data logging. The estimation system is designed linear, which does not consider affinity and other hardcore computer vision theories, in the trade-off of a considerable effort to setup for each data logging session. The first of its nature, a detailed study is performed on the chile pepper plant and fruit to study the feasibility and challenges for robotic harvesting of chile pepper, which is immensely needed for researchers who are contributing in this field. Based on the collected data, it was shown that an optimum location of the robotic arm can be extracted. For Braccio robot arm, the optimum location was found to be ( 180 , 0 , 120 ) , with respect to the world frame. The forward and inverse kinematics of Braccio robot was derived for workspace analysis and motion planning. A chile pepper harvesting robot is developed with a customized cutting mechanism, in which its performance was examined in a series of in-laboratory harvesting tests. The developed harvesting robot showed promising results with localization success rate, detachment success rate, harvest success rate, damage rate, and cycle time of 37.7%, 65.5%, 24.7%, 6.9%, and 7 s, respectively. With this study, it is shown that the chile pepper harvesting is feasible with robotics, and, in the future, more efforts should be put into this area by the researchers. By eliminating the inherent drawbacks and limitations in the presented system, the success rate can be increased.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/robotics10030094/s1, Table S1: Harvesting results, Table S2: Chile peppers’ 3D location, Video S1: Harvesting supplementary videos.

Author Contributions

Conceptualization, M.U.M. and M.H.-J.; methodology, M.U.M. and M.H.-J.; software, M.U.M. and M.H.-J.; validation, M.U.M.; formal analysis, M.U.M.; investigation, M.U.M.; resources, M.H.-J.; data curation, M.U.M.; writing—original draft preparation, M.U.M. and M.H.-J.; writing—review and editing, M.H.-J.; visualization, M.U.M. and M.H.-J.; supervision, M.H.-J.; funding acquisition, M.H.-J. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by a grant (award number 2021-67021-34203) from USDA/NIFA and a seed funding from the NMSU Center of Excellence in Sustainable Food and Agricultural Systems (CESFAS) program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the Supplementary Materials provided in this article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DoFDegrees of freedom
D-HDenavit–Hartenberg
DCDirect current
CADComputer aided design
SLAStereolithography
3DThree-dimensional
SDStandard deviation
SDKSoftware development kit
RGBRed-Green-Blue

References

  1. Bosland, P.W.; Bailey, A.L.; Cotter, D.J. Growing Chiles in New Mexico; Guide H-New Mexico State University, Cooperative Extension Service (USA): Las Cruces, NM, USA, 1991. [Google Scholar]
  2. Bosland, P.W.; Walker, S. Growing Chiles in New Mexico [Guide H-230]; Guide H-New Mexico State University, Cooperative Extension Service (USA): Las Cruces, NM, USA, 2014. [Google Scholar]
  3. Taylor, J.E.; Charlton, D.; Yúnez-Naude, A. The end of farm labor abundance. Appl. Econ. Perspect. Policy 2012, 34, 587–598. [Google Scholar] [CrossRef]
  4. Hertz, T.; Zahniser, S. Is there a farm labor shortage? Am. J. Agric. Econ. 2013, 95, 476–481. [Google Scholar] [CrossRef] [Green Version]
  5. 2020 New Mexico Chile Production; Technical Report; USDA National Agriculture Statistics Service: Washington, DC, USA, 2021.
  6. Russo, V.M. Peppers: Botany, Production and Uses; CABI Publishing: Wallingford, UK, 2012; pp. 1–280. [Google Scholar] [CrossRef]
  7. Miles, J.; Hinz, W.; Pike, W. Development of a Mechanism for Picking Chile Peppers. Trans. ASAE 1978, 21, 419–0421. [Google Scholar] [CrossRef]
  8. Funk, P.A.; Walker, S.J.; Herbon, R.P. A systems approach to chile harvest mechanization. Int. J. Veg. Sci. 2011, 17, 296–309. [Google Scholar] [CrossRef]
  9. Funk, P.A.; Walker, S.J. Evaluation of Five Green Chile Cultivars Utilizing Five Different Harvest Mechanisms. Appl. Eng. Agric. 2010, 26, 955–964. [Google Scholar] [CrossRef]
  10. Bac, C.W.; Van Henten, E.J.; Hemming, J.; Edan, Y. Harvesting Robots for High-value Crops: State-of-the-art Review and Challenges Ahead. J. Field Robot. 2014, 31, 888–911. [Google Scholar] [CrossRef]
  11. Bechar, A.; Vigneault, C. Agricultural robots for field operations: Concepts and components. Biosyst. Eng. 2016, 149, 94–111. [Google Scholar] [CrossRef]
  12. Bonadies, S.; Lefcourt, A.; Gadsden, S.A. A survey of unmanned ground vehicles with applications to agricultural and environmental sensing. Autonomous air and ground sensing systems for agricultural optimization and phenotyping. Int. Soc. Opt. Photonics 2016, 9866, 98660Q. [Google Scholar]
  13. Zhao, D.-A.; Lv, J.; Ji, W.; Zhang, Y.; Chen, Y. Design and control of an apple harvesting robot. Biosyst. Eng. 2011, 110, 112–122. [Google Scholar]
  14. Feng, Q.; Zou, W.; Fan, P.; Zhang, C.; Wang, X.; Xiong, Y.; From, P.J.; Isler, V.; Davidson, J.R.; Bulanon, D.M.; et al. Fruit detection system and an end effector for robotic harvesting of Fuji apples. Agric. Eng. Int. CIGR J. 2019, 12, 203–210. [Google Scholar] [CrossRef]
  15. Liu, T.H.; Zeng, X.R.; Ke, Z.H. Design and prototyping a harvester for litchi picking. In Proceedings of the 2011 Fourth International Conference on Intelligent Computation Technology and Automation, Shenzhen, China, 28–29 March 2011; Volume 2, pp. 39–42. [Google Scholar] [CrossRef]
  16. Baeten, J.; Donné, K.; Boedrij, S.; Beckers, W.; Claesen, E. Autonomous fruit picking machine: A robotic apple harvester. Springer Tracts Adv. Robot. 2008, 42, 531–539. [Google Scholar] [CrossRef] [Green Version]
  17. Tanigaki, K.; Fujiura, T.; Akase, A.; Imagawa, J. Cherry-harvesting robot. Comput. Electron. Agric. 2008, 63, 65–72. [Google Scholar] [CrossRef]
  18. Aljanobi, A.A.; Al-Hamed, S.A.; Al-Suhaibani, S.A. A Setup of Mobile Robotic Unit for Fruit Harvesting. In Proceedings of the 19th International Workshop on Robotics in Alpe-Adria-Danube Region, RAAD 2010-Proceedings, Budapest, Hungary, 24–26 June 2010; pp. 105–108. [Google Scholar] [CrossRef]
  19. Hayashi, S.; Ganno, K.; Ishii, Y.; Tanaka, I. Robotic Harvesting System for Eggplants. Jpn. Agric. Res. Q. JARQ 2002, 36, 163–168. [Google Scholar] [CrossRef] [Green Version]
  20. Davidson, J.; Bhusal, S.; Mo, C.; Karkee, M.; Zhang, Q. Robotic Manipulation for Specialty Crop Harvesting: A Review of Manipulator and End-Effector Technologies. Glob. J. Agric. Allied Sci. 2020, 2, 25–41. [Google Scholar] [CrossRef]
  21. Gunderman, A.L.; Collins, J.; Myer, A.; Threlfall, R.; Chen, Y. Tendon-Driven Soft Robotic Gripper for Berry Harvesting. arXiv 2021, arXiv:2103.04270. [Google Scholar]
  22. Ceccarelli, M.; Figliolini, G.; Ottaviano, E.; Mata, A.S.; Criado, E.J. Designing a robotic gripper for harvesting horticulture products. Robotica 2000, 18, 105–111. [Google Scholar] [CrossRef] [Green Version]
  23. Kultongkham, A.; Kumnon, S.; Thintawornkul, T.; Chanthsopeephan, T. The design of a force feedback soft gripper for tomato harvesting. J. Agric. Eng. 2021, 52. [Google Scholar] [CrossRef]
  24. Hohimer, C.J.; Wang, H.; Bhusal, S.; Miller, J.; Mo, C.; Karkee, M. Design and field evaluation of a robotic apple harvesting system with a 3d-printed soft-robotic end-effector. Trans. ASABE 2019, 62, 405–414. [Google Scholar] [CrossRef]
  25. Bulanon, D.M.; Kataoka, T. Fruit Detection System and an End Effector for Robotic Harvesting of Fuji Apples; American Society of Agricultural and Biological Engineers: St. Joseph, MI, USA, 2010. [Google Scholar]
  26. Van Henten, E.J.; Hemming, J.; Van Tuijl, B.; Kornet, J.; Meuleman, J.; Bontsema, J.; Van Os, E. An autonomous robot for harvesting cucumbers in greenhouses. Auton. Robot. 2002, 13, 241–258. [Google Scholar] [CrossRef]
  27. Van Henten, E.; Van’t Slot, D.; Hol, C.; Van Willigenburg, L. Optimal manipulator design for a cucumber harvesting robot. Comput. Electron. Agric. 2009, 65, 247–257. [Google Scholar] [CrossRef]
  28. Liu, J.; Li, P.; Li, Z. A multi-sensory end-effector for spherical fruit harvesting robot. In Proceedings of the IEEE International Conference on Automation and Logistics, ICAL 2007, Jinan, China, 18–21 August 2007; pp. 258–262. [Google Scholar] [CrossRef]
  29. Jia, B.; Zhu, A.; Yang, S.X.; Mittal, G.S. Integrated gripper and cutter in a mobile robotic system for harvesting greenhouse products. In Proceedings of the 2009 IEEE International Conference on Robotics and Biomimetics, ROBIO 2009, Guilin, China, 19–23 December 2009; pp. 1778–1783. [Google Scholar] [CrossRef]
  30. Reed, J.; Miles, S.; Butler, J.; Baldwin, M.; Noble, R. AE—Automation and emerging technologies: Automatic mushroom harvester development. J. Agric. Eng. Res. 2001, 78, 15–23. [Google Scholar] [CrossRef]
  31. Foglia, M.M.; Reina, G. Agricultural robot for radicchio harvesting. J. Field Robot. 2006, 23, 363–377. [Google Scholar] [CrossRef]
  32. Sakai, S.; Iida, M.; Osuka, K.; Umeda, M. Design and control of a heavy material handling manipulator for agricultural robots. Auton. Robot. 2008, 25, 189–204. [Google Scholar] [CrossRef] [Green Version]
  33. Wei, J.; Yi, D.; Bo, X.; Guangyu, C.; Dean, Z. Adaptive variable parameter impedance control for apple harvesting robot compliant picking. Complexity 2020. [Google Scholar] [CrossRef]
  34. 3D Printing Materials: Engineering Resins|Formlabs. Available online: https://formlabs.com/ (accessed on 21 July 2021).
  35. Moghimi, A.; Aghkhani, M.H.; Golzarian, M.R.; Rohani, A.; Yang, C. A robo-vision algorithm for automatic harvesting of green bell pepper. In Proceedings of the 2015 ASABE Annual International Meeting, American Society of Agricultural and Biological Engineers, New Orleans, LA, USA, 26–29 July 2015; p. 1. [Google Scholar]
  36. Kitamura, S.; Oka, K. Recognition and cutting system of sweet pepper for picking robot in greenhouse horticulture. In Proceedings of the IEEE International Conference Mechatronics and Automation, Niagara Falls, ON, Canada, 29 July–1 August 2005; Volume 4, pp. 1807–1812. [Google Scholar]
  37. Yuan, T.; Li, W.; Feng, Q.; Zhang, J. Spectral Imaging for Greenhouse Cucumber Fruit Detection Based on Binocular Stereovision; American Society of Agricultural and Biological Engineers: Pittsburgh, PA, USA, 2010; p. 1. [Google Scholar]
Figure 1. Harvesting scheme showing all the components.
Figure 1. Harvesting scheme showing all the components.
Robotics 10 00094 g001
Figure 2. Lab setup for testing of harvesting robot.
Figure 2. Lab setup for testing of harvesting robot.
Robotics 10 00094 g002
Figure 3. The image that the operator is seeing and clicks on the ripe chile pepper.
Figure 3. The image that the operator is seeing and clicks on the ripe chile pepper.
Robotics 10 00094 g003
Figure 4. Braccio robot arm.
Figure 4. Braccio robot arm.
Robotics 10 00094 g004
Figure 6. Geometric representation of the Braccio robot arm for inverse kinematics.
Figure 6. Geometric representation of the Braccio robot arm for inverse kinematics.
Robotics 10 00094 g006
Figure 7. (a) The workspace in magenta, dexterous space for β = 0 ° in blue, and dexterous space for β = ± 25 ° in green, shown in a planar view for a specific base rotation (b) shown in 3D point cloud data for better visualization.
Figure 7. (a) The workspace in magenta, dexterous space for β = 0 ° in blue, and dexterous space for β = ± 25 ° in green, shown in a planar view for a specific base rotation (b) shown in 3D point cloud data for better visualization.
Robotics 10 00094 g007
Figure 8. (a) CAD model of the cutting mechanism. (b) Fabricated cutter installed on the robot arm.
Figure 8. (a) CAD model of the cutting mechanism. (b) Fabricated cutter installed on the robot arm.
Robotics 10 00094 g008
Figure 9. Actual camera and table setup shown with frames of references.
Figure 9. Actual camera and table setup shown with frames of references.
Robotics 10 00094 g009
Figure 10. Dimensions of chile pepper taken into consideration in size data collection.
Figure 10. Dimensions of chile pepper taken into consideration in size data collection.
Robotics 10 00094 g010
Figure 11. Histogram of the collected data for (a) diameter of chile pepper, (b) length of chile pepper, (c) diameter of the stem, and (d) available length of the stem.
Figure 11. Histogram of the collected data for (a) diameter of chile pepper, (b) length of chile pepper, (c) diameter of the stem, and (d) available length of the stem.
Robotics 10 00094 g011
Figure 12. (a) Spatial distribution shown on physical plant, (b) chile distribution shown with respect to the center of the bottom face of the chile pot, and (c) optimum location of the robot to have maximum chile peppers in the dexterous space.
Figure 12. (a) Spatial distribution shown on physical plant, (b) chile distribution shown with respect to the center of the bottom face of the chile pot, and (c) optimum location of the robot to have maximum chile peppers in the dexterous space.
Robotics 10 00094 g012
Figure 13. Algorithm for the harvesting program.
Figure 13. Algorithm for the harvesting program.
Robotics 10 00094 g013
Figure 14. (a) Actuator trajectory for simulation. (b) MATLAB™ simulation of Braccio with joint angles.
Figure 14. (a) Actuator trajectory for simulation. (b) MATLAB™ simulation of Braccio with joint angles.
Robotics 10 00094 g014
Figure 15. Challenging cases for robotic harvesting. (a) Chile pepper with almost zero stem available for cutting. (b) A cluster of chile pepper in the same vicinity. (c) A chile pepper with its stem very close of main or sub-branch of the plant.
Figure 15. Challenging cases for robotic harvesting. (a) Chile pepper with almost zero stem available for cutting. (b) A cluster of chile pepper in the same vicinity. (c) A chile pepper with its stem very close of main or sub-branch of the plant.
Robotics 10 00094 g015
Figure 16. MATLAB simulation (x-y plane view) of showing chile peppers which are not reachable (in the bounding box) because of the plant stem in the pathway as an obstacle.
Figure 16. MATLAB simulation (x-y plane view) of showing chile peppers which are not reachable (in the bounding box) because of the plant stem in the pathway as an obstacle.
Robotics 10 00094 g016
Figure 17. Harvesting operation of the robot shown in frames.
Figure 17. Harvesting operation of the robot shown in frames.
Robotics 10 00094 g017
Table 2. Accuracy and precision results for Braccio robotic arm.
Table 2. Accuracy and precision results for Braccio robotic arm.
x-axisy-axisz-axis
μ −5.7−2.41.7
σ 1 0.91.00.9
σ 2 0.30.60.7
σ 3 0.70.91.0
All dimensions in mm.
Table 3. Error (mm) in 3D location estimation.
Table 3. Error (mm) in 3D location estimation.
x-axisy-axisz-axis
Mean−1.05.2−2.5
SD2.65.71.6
Table 4. Variation in chile fruit dimensions and anatomical features.
Table 4. Variation in chile fruit dimensions and anatomical features.
DdLl
Mean29.375.65113.3822.14
SD5.350.8621.1012.20
All dimensions in mm.
Table 5. Results of lab testing of robotic chile pepper harvesting.
Table 5. Results of lab testing of robotic chile pepper harvesting.
Testing ParameterResultUnits
Localization Success37.7%
Detachment Success65.5%
Harvest Success24.7%
Cycle Time7s
Damage Rate6.9%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Masood, M.U.; Haghshenas-Jaryani, M. A Study on the Feasibility of Robotic Harvesting for Chile Pepper. Robotics 2021, 10, 94. https://doi.org/10.3390/robotics10030094

AMA Style

Masood MU, Haghshenas-Jaryani M. A Study on the Feasibility of Robotic Harvesting for Chile Pepper. Robotics. 2021; 10(3):94. https://doi.org/10.3390/robotics10030094

Chicago/Turabian Style

Masood, Muhammad Umar, and Mahdi Haghshenas-Jaryani. 2021. "A Study on the Feasibility of Robotic Harvesting for Chile Pepper" Robotics 10, no. 3: 94. https://doi.org/10.3390/robotics10030094

APA Style

Masood, M. U., & Haghshenas-Jaryani, M. (2021). A Study on the Feasibility of Robotic Harvesting for Chile Pepper. Robotics, 10(3), 94. https://doi.org/10.3390/robotics10030094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop