Next Article in Journal
A Study on Robust Finite-Time Visual Servoing with a Gyro-Stabilized Surveillance System
Previous Article in Journal
Vibration Control of Car Body and Wheel Motions for In-Wheel Motor Vehicles Using Road Type Classification
Previous Article in Special Issue
Optimization Design and Experiment of High-Speed Drag-Reducing Trencher Based on Conservation Tillage
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Point Recognition for Seedlings Using Kernel Density Estimators and Pyramid Histogram of Oriented Gradients

by
Moteaal Asadi Shirzi
*,† and
Mehrdad R. Kermani
Department of Electrical and Computer Engineering, University of Western Ontario, London, ON N6A 3K7, Canada
*
Author to whom correspondence should be addressed.
Current address: Department of Electrical and Computer Engineering, ACEB 3400, Western University, London, ON N6A 5B9, Canada.
Actuators 2024, 13(3), 81; https://doi.org/10.3390/act13030081
Submission received: 25 January 2024 / Revised: 18 February 2024 / Accepted: 18 February 2024 / Published: 21 February 2024
(This article belongs to the Special Issue Design and Control of Agricultural Robotics)

Abstract

:
This paper introduces a new real-time method based on a combination of kernel density estimators and pyramid histogram of oriented gradients for identifying a point of interest along the stem of seedlings suitable for stem–stake coupling, also known as the ‘clipping point’. The recognition of a clipping point is a required step for automating the stem–stake coupling task, also known as the clipping task, using the robotic system under development. At present, the completion of this task depends on the expertise of skilled individuals that perform manual clipping. The robotic stem–stake coupling system is designed to emulate human perception (in vision and cognition) for identifying the clipping points and to replicate human motor skills (in dexterity of manipulation) for attaching the clip to the stem at the identified clipping point. The system is expected to clip various types of vegetables, namely peppers, tomatoes, and cucumbers. Our proposed methodology will serve as a framework for automatic analysis and the understanding of the images of seedlings for identifying a suitable clipping point. The proposed algorithm is evaluated using real-world image data from propagation facilities and greenhouses, and the results are verified by expert farmers indicating satisfactory performance. The precise outcomes obtained through this identification method facilitate the execution of other autonomous functions essential in precision agriculture and horticulture.

Graphical Abstract

1. Introduction

One of the most time-consuming and laborious tasks in facilities specializing in seedling propagation is the attachment of a clip at a specific location along the seedling stem to couple the stem to a wooden stake. This action, called stem–stake coupling or clipping, requires a human worker to bend over a wide tray containing the seedlings or to place a knee on the floor (as shown in Figure 1) and use two hands to put a plastic clip around the seedling’s main stem and the wooden stake. The clip provides support for the seedling during transportation and further growth. In greenhouses, such clips are used to force the plant to grow in the desired direction and to expedite growth and streamline harvesting. The clips are also necessary for supporting the plant’s heavy weight during growth and fruiting. The clipping task involves a significant amount of manual labor to process a large number of seedlings in a vast arena. This is a physically demanding and painstaking task. In a typical propagation facility (e.g., Roelands Plant Farms in Lambton Shores, ON, Canada), approximately 25 million seedlings are germinated annually. Each seedling is manually clipped before shipping. The sheer volume of seedlings processed each year underscores the necessity and advantages of developing a robotic solution for automated clipping. Such a solution would lead to tangible improvements in process efficiency, productivity, and product quality. Additionally, it would significantly reduce the labor costs associated with the clipping task and prevent work-related injuries, such as back injuries, strains, and sprains resulting from performing awkward body positions for an extend period of time.
A robotic clipping solution requires two main components: a machine vision unit that analyzes an image of a seedling and identifies the clipping point and a mechatronic unit that performs the act of clipping. The focus of this paper is on the machine vision unit that replicates human visual and cognitive functionalities to understand an image and identify a suitable clipping point along the seedling’s main stem for different types of vegetables, namely peppers, tomatoes, and cucumbers. In recent years, machine vision and various visual processing algorithms have seen significant advancements in agriculture-related applications [1]. These applications span a wide range of tasks, including spraying fertilizers and pesticides [2], plant detection and harvesting [3], grafting, irrigation, automatic grading [4], plant disease detection [5], ripe fruit identification [6], crop-weed classification [7], and cutting or pruning [8].
Real-time point recognition refers to the ability of a system or algorithm to identify and locate a specific point or key points in an image or video stream in real time. Real-time recognition of a suitable clipping point is a challenging problem in image processing as it requires the modeling of the human’s cognitive processes and past experiences. We extensively reviewed various approaches in the literature that could potentially address this problem. Each approach has its strengths and limitations when applied to this specific task. Some studies, such as those examining multi-feature patch-based segmentation [9], adaptive spectral–spatial gradient sparse regularization [10], the adaptive snake algorithm model [11], kernel-based algorithms based on the adaptive Kalman filter [12], and the SIFT feature descriptor with hybrid optimization algorithms [13], offer rigorous mathematical analysis and quantitative measures of computer vision. These algorithms encounter significant challenges in dealing with occlusions caused by leaves and neighboring seedlings, varying lighting conditions, and different backgrounds in various greenhouse settings. Consequently, adapting vision algorithms for real-time and reliable results in such diverse conditions becomes crucial. Alternatively, end-to-end deep learning models [14,15] are particularly effective in scenarios where the input data can be directly mapped to the desired output without the need for extensive pre-processing or feature engineering. These models require hundreds, if not thousands, of labeled images for training the network for each seedling type, leading to substantial processing times, also known as latency or inference time [16]. These challenges render the abovementioned algorithms less suitable and less robust for our specific application. Recently, feature-based methods employing machine learning techniques and complex optimization frameworks [17,18] have shown promise in modeling processes that rely on heuristic knowledge and human experience. However, feature-based methods often require manually engineered features which limit their performance to specific conditions or scenarios for which these features were designed. It can be challenging to create features that adequately represent the underlying patterns in all possible conditions. This process becomes increasingly difficult as the complexity and variability of conditions grow considering seedling types, sizes, shapes, and patterns. In addition, complex optimization frameworks can be computationally intensive and may require substantial resources, making them impractical for real-time applications or situations with limited computational power [19].
To address these challenges, we take advantage of a hybrid approach based on analytical image processing and data-driven learning algorithms. By integrating traditional and modern approaches, we aim to overcome the problems associated with varying lighting and environmental conditions and adapting to different types of seedlings. The analytical part of our proposed method calculates feature descriptors based on point density variation, kernel density estimation, and pyramid histogram of oriented gradients. The data-driven learning part of our algorithm uses Extreme Gradient Boosting to select the most suitable clipping point using the feature descriptors obtained in the analytical part. We applied the algorithm to one hundred images of real seedlings from “Roelands Plant Farms” and expert farmers reviewed the results to evaluate the performance and accuracy of the proposed method. We equipped a robotic arm with a proprietary clipping device and a stereo camera to assess the performance of our methodology. The accurate results derived from this identification method open up the possibility of implementing other autonomous functionalities required in precision agriculture.
The contributions of this work are as follows:
-
Introducing a real-time hybrid method to find the most suitable clipping point on seedlings.
-
Comparing proposed method with deep learning based methods.
-
Real-time evaluation of the results using a large number of in situ images.
-
Experimental studies using a robotic system to validate the performance of the proposed method
The organization of the paper is as follows: Section 2 describes the materials and methods. Section 3 contains the evaluation strategy and dataset. Section 4 represents the results, and Section 5 summarizes and concludes the paper.

2. Materials and Methods

2.1. Suitable Clipping Point

The selection of a clipping point for each seedling varies based on factors such as type, age, height, dimensions of inter-nodes and petioles, size and arrangement of leaves, and the number of nodes. Farmers put the clip on the highest point, higher than the uppermost node on the main stem. However, if the length of the main stem is too short between the two nodes or the seedling height surpasses that of the stake, the clipping point is selected below the highest node. The leaves around the highest node are often dense and some parts of the main stem are behind the leaves. The thickness and shape of the stem and petioles are the same, especially in the top parts of a seedling. Thus, recognizing the main stem from the petioles using machine vision is not an easy task. The recognition process also depends on the shape and type of seedlings. The spatial distance between the stake and stem and the accessibility of the point through the leaves must be taken into consideration, which further complicates the selection of a correct point for clipping. The selection of the clipping point is not a straight forward task but rather a cognitive process that relies heavily on heuristic information. Figure 2 shows some seedlings and the preferred clipping points that expert farmers have validated. As seen, the locations of the selected clipping points do not follow a specific pattern that can be algorithmized.

2.2. General Concept

The recognition of a suitable clipping point using the combination of a kernel density estimator (KDE) and pyramid histogram of oriented gradients (PHOG) is a promising approach in computer vision and image processing. The KDE is a non-parametric way to estimate the probability density function of a random variable, while PHOG is a feature descriptor that captures the local shape information in an image. Figure 3 shows the schematic of our proposed method using images taken with a stereo camera. In the pre-processing step, after camera calibration using Zhang’s method [20], the global contrast of the image is enhanced using the histogram equalization. To normalize all images for different lighting conditions, the cumulative distribution function matching method is used to balance the target images with a reference image by removing systematic differences or re-scaling the histogram of images taken in different lighting conditions [21]. High-boost filters and morphological boundary enhancement [22] are used to enhance and restore images before the object detection module. In the next step, the plant, stem, and stake are detected. We then use the boundary and skeleton of the segmented plant to find the region of interest (ROI), limit the searching area, avoid unwanted points, and accelerate the process.
After the region of interest is selected, the point density variation (PDV) measures the disparity in each area relative to the color intensity of the plant. Using PDV allows us to extend the characteristics of each area of the image to the adjacent areas. As a result, the algorithm not only focuses on specific areas of the image, but it considers the characteristics of other pixels around the area. It improves the capabilities of the algorithm to determine some areas of the image with the same pattern but with different environments around them.
In the next step, a KDE is applied to estimate the probability density function (PDF) of key point locations. This process aids in understanding the spatial distribution of key points across the image. The KDE works by smoothing raw data to estimate the population density of such finite data samples. In image processing, the KDE is particularly useful for estimating the underlying continuous distribution of data, whether it be pixel intensities or feature values. The resulting PDF provides a representation of how these values are distributed throughout the image.
Next, we construct a feature vector using the PHOG method. This method offers a rich representation by capturing local texture and shape information within the image. It achieves this by constructing histograms of gradient orientations at varying scales and positions, arranged in a pyramid-like structure. This approach captures both local and global image characteristics.
Following this, the XGBoost machine learning algorithm is applied. It utilizes both the PDF, derived from the KDE, and the feature vector created using the PHOG and PVD methods. These elements are combined to analyze the region of interest (ROI). XGBoost then identifies the most suitable clipping point based on this comprehensive analysis.
A stereo camera system is used for stereo triangulation to calculate the real-world 3D coordinates of the suggested clipping point. In the last step, the algorithm checks the accessibility of the wooden stake and stem for the robotic arm and clipping device. If suitable, it maps the coordinate system of the stereo camera to that of the robot, providing essential sensory feedback for the robot’s controller.

2.2.1. Plant, Stem, and Stake Detection

The quality of the plant recognition is crucial for having accepted results in subsequent steps. After an exhaustive review of existing automatic segmentation algorithms, which included both classic image processing and deep learning approaches, we assessed the advantages and disadvantages of these methods. Consequently, we opted for an adaptive feature-based plant segmentation algorithm [23] as it provides the required accuracy for effective plant detection. This method is powerful since it can deal with various seedlings with different lighting conditions and complex backgrounds. Noise is removed from the segmented plant using morphological filtering techniques [24] after plant recognition. We used a combination of morphological techniques, such as hit-and-miss, thinning-and-thickening, convex hull, and morphological gradient to eliminate the leaves from the image [25]. The effectiveness of morphological operations depends on the specific characteristics of the image and the nature of the problem we are trying to solve. By choosing the right structuring element, morphological processing may remove special parts of the image based on the size and shape of these parts. In order to identify parts of the stem that are hidden behind the leaf or wrongly removed from the image when using the mathematical morphology operators, a fifth-degree polynomial spline was chosen to estimate some hidden parts of the stem. The main premise of using this technique is that the stem has a smooth shape with no sudden changes in the shape or its angle. A spline matching algorithm and hybridization of the Otsu method and median filter detects the wooden stake [26]. The wooden stake is almost vertically straight. Thus, hidden and covered parts of the stake can be detected using simple partial spline matching. Given that the stereo camera captures images from varying distances, images are resized and cropped to mitigate the effects of distance variation. The distance of the stem is calculated from the stereo matching algorithm [27]. The size ratio k is computed as follows:
k = δ D n B f
where  δ  represents the disparity,  D n  is the normal distance, B is the baseline distance between the cameras, and f is the focal length of the stereo camera.

2.2.2. Region of Interest

To avoid unwanted points and accelerate the process of finding the most suitable clipping point, the proposed real-time point recognition method uses the boundary and skeleton of the seedling to find the region of interest (ROI) and limit the search area. The borders of the ROI are computed as follows:
S ¯ i i = 1 n ( P r ( i ) S ( i ) ) ( n ϕ r ) < P i < S ¯ i + i = 1 n ( P l ( i ) S ( i ) ) ( n ϕ l )
and
S ¯ j j = 1 m ( P t ( j ) S ( j ) ) ( m ϕ t ) < P j < S ¯ j
where  S ¯ i  and  S ¯ j  are the mean values of the skeleton in the x and y directions, respectively, for all non-null pixels;  P r P l , and  P t  are, respectively, the right, left, and top values of the seedling boundaries for non-null values;  S ( i )  and  S ( j )  are the values of the skeleton in pixel  ( i , j ) ; and  ϕ r ϕ l , and  ϕ t  are the number of null values for the right, left, and top of the boundary of the plant, respectively.

2.2.3. Point Density Variation

The point density variation illustrates the disparity in color intensity across a map [28]. For each color channel i, a Gaussian mixture model represents the distribution as follows:
P i ( x ) = k = 1 K π i k N i ( x | μ i k , Σ i k )
where  π i k  are the mixing coefficients satisfying
k = 1 K π i k = 1 , π i k 0
The  P i ( x )  is the combined Gaussian distribution of intensities for color channel i, in that  N ( x | μ i k , Σ i k )  is the Gaussian density with mean  μ i k  and covariance  Σ i k . Assuming diagonal covariance matrices of the form  Σ i = σ i 2  [29], the Gaussian model for each color channel simplifies to
P i ( x ) = 1 ( 2 π σ i 2 ) 1 2 e | | x k μ i | | 2 2 σ i 2
where  μ i  and  Σ i  are the mean and covariance values, respectively.
We applied this Gaussian model to each color channel of the segmented image. The point density variation of the seedling is represented by combining these Gaussian distributions, weighted by their respective mixing coefficients.

2.2.4. Kernel Density Estimator

In most computer vision and pattern recognition applications, feature spaces are complex and often contain noise, making them difficult to describe effectively using common parametric models. Instead, non-parametric density estimation techniques are widely used to analyze arbitrarily structured feature spaces [30]. The kernel density estimator (KDE), as a non-parametric density estimation technique, calculates the feature density in a neighborhood around those features. The density function is estimated by summing kernel functions (typically Gaussians) centered at the data points [31].
The comparison between histogram and kernel density estimation shows that a bandwidth associated with the kernel function controls the smoothness of the estimated densities. More data points allow for a narrower bandwidth, leading to a more accurate density estimate. The kernel density estimator distributes the known population quantity for each point across the location of random non-parametric variables [32]. Consequently, point density variation estimates the density of stem intensity, making kernel density estimation a fundamental data smoothing technique that makes inferences about the population from a finite data sample [33]. Assuming  X 1 , X 2 , , X n  are independent and identically distributed points that have a univariate distribution with an unknown density at any given point x, the kernel density estimator is defined as
f ¯ ( x ) = 1 n i = 1 h K h ( x x i ) = 1 n h i = 1 h K ( x x i h )
where K is the kernel,  h > 0  is a smoothing parameter called the bandwidth, and  K h ( x ) = 1 h × K ( x h )  is the scaled kernel. As a rule of thumb [34], if the Gaussian basis function is used to approximate univariate data, the optimal choice for the bandwidth is
h = 0.9 min ( σ ˜ , I Q R 1.34 ) n 1 n
where  σ ˜  is the approximate standard deviation and  I Q R  is the interquartile range.

2.2.5. Pyramid Histogram of Oriented Gradients

The pyramid histogram of oriented gradients (PHOG) is a feature descriptor used in computer vision for image analysis and object recognition [35]. It was introduced as an extension of histogram of oriented gradients (HOG) to capture both local and global information about the shape and structure of objects in an image. HOG represents a signature of the features of spatial regions [36] and is particularly useful when dealing with images that contain objects at different scales. PHOG divides an image into small cells and computes HOG within each cell. These histograms provide a representation of the local texture and edge information. PHOG extends HOG by incorporating a pyramid structure. Instead of analyzing the entire image at a single scale, PHOG divides the image into multiple scales or levels. This helps in capturing information at different resolutions, allowing the descriptor to be more robust to variations in object size and scale. At each level of the pyramid, HOG features are computed for local image regions (cells). The features are then concatenated to form a representation for that level. Additionally, these level-specific representations are combined across the pyramid to capture global information. For computing the HOG of each cell, we used the first quarter of the cell to compute the histogram gradient in sixteen directions. We considered all different directions and the number of pixels within each direction. These steps were repeated for other quarters. The resulting orientation histograms were concatenated to create a long histogram gradient (LHG). Figure 4 shows the main steps of computing the HOG.
To make the descriptor robust to changes in illumination and contrast, normalization is often applied. The largest gradient direction was selected as the principal orientation. We normalized the principal orientation of all histogram gradients of the cells. Normalizing the histogram diminishes the effect of scaling and ignores the magnitude of the gradient. The rotation ignores the effect of different orientations. We used (10) and (11) to calculate the size and angle of the principle of the LHG:
θ = t a n 1 ( I y | m a x / I x | m a x )
d = ( I y | m a x ) 2 + ( I x | m a x ) 2
where  I x  is the rate of the change of the LHG in x direction and  I y  is the rate of the change of the LHG in y direction. This involves dividing the feature vectors by a local or global measure of their magnitude. The final PHOG descriptor was formed by concatenating the feature vectors from different levels of the pyramid. This results in a comprehensive representation of the image structure at multiple scales.

2.2.6. Extreme Gradient Boosting

Extreme Gradient Boosting (XGBoost) is primarily designed for supervised learning tasks. It is well equipped to address non-linear relationships between input features and target variables [37] and can capture intricate patterns and interactions within high-dimensional feature spaces [38]. XGBoost belongs to the ensemble learning category, specifically to the gradient boosting framework. Ensemble learning involves combining the predictions of multiple learners to create a strong learner. Gradient boosting is a technique that builds a series of models sequentially, with each new model attempting to correct the errors of the combined ensemble of models built so far. This is achieved by constructing a collection of decision trees, where each tree can represent various facets of the nonlinear relationships present in the data. The algorithm works by minimizing the cumulative loss over the entire ensemble of models, where each new model is added in a way that reduces the overall error. XGBoost is known for its efficiency and scalability, making it suitable for a wide range of machine learning tasks.
To reduce redundant information in the hyperspectral data, feature selection based on the importance of boosting was used. The implementation of model development was programmed using the Scikit-learn, XGboost, and LightGBM packages in Python 3.9 [39].

2.2.7. Overall Approach

Real-time point recognition using kernel density estimators and the pyramid histogram of oriented gradients (KDE-PHOG) algorithm follows multiple steps to find the clipping point. In this section, we briefly describe these steps using an example. In Figure 5, the plant, stem, and stake are detected following camera calibration and pre-processing. This detection uses adaptive feature-based plant segmentation, morphological techniques, and a hybrid approach combining the Otsu method and a median filter. Subsequently, the plant’s boundary and skeleton are obtained. The region of interest (ROI) is then defined based on these boundary and skeleton features of the plant.
We used PHOG to select multiple cells with different scales and levels in a pyramid structure to capture information at different resolutions. Figure 6 shows samples of the cells and histogram of PDV, the feature vector of the KDE, and normalized PHOG. It is clear that the PDF of the cell that contains the suitable clipping point is unique and different from the other cells, and XGBoost can recognize such a cell.
In the subsequent step, the ideal point for clipping is identified within the selected cell scene. This selection takes into account the presence of both the stake and the stem in that specific area as well as the distance between them. Additionally, the algorithm checks for a corresponding point on the stake near the selected clipping point. If such a point is accessible, the KDE-PHOG method is utilized to calculate the distance and depth of the point using images from a stereo camera. In the final step, the real coordinates of the clipping point are transmitted to the robotic arm, which then positions the clipping device at the specified point. Figure 7 illustrates these steps.

3. Evaluation Strategy and Datasets

To evaluate the precision of the proposed algorithm in identifying suitable clipping points, we conducted a comparative analysis against three other automatic techniques, including deep learning approaches based on Convolutional Neural Networks (CNNs). Table 1 provides a brief description of these three methods.
To compare our proposed method with others, we calculated the mean Average Precision (mAP) to evaluate the precision of clipping point detection.
Average Precision (AP) is computed by calculating the area under the precision–recall curve. It shows the relationship between two important evaluation metrics: precision (how many selected items are relevant) and recall (how many relevant items are selected). Recall is defined as follows:
R e c a l l = T P T P + F N
and precision is defined as:
P r e c i s i o n = T P T P + F P
where  T P  represents number of correctly detected points,  F P  is the number of miss-detection points, and  F N  is the number of false alarms. Mean Average Precision,  m A P , is defined as:
m P A = 0 1 P ( R ) . P d R
where P is precision, R is recall, and  m P A  measures the performance of the detector. In the case of a single class, the mean Average Precision ( m A P ) is equivalent to the Average Precision ( A P ) for that class, and it can be used as a metric to standardize performance measurement.  A P  represents the area under the precision–recall curve, and the higher the  A P  value is, the better the detection accuracy is.
Our dataset contained real images of five different seedling types, including Beit Alpha cucumber, tomato, chili pepper, bell pepper, and national pickling cucumber. The images had different resolutions and included multi-scale seedlings with different lighting conditions and backgrounds. We utilized a dataset of 500 sample images paired with the corresponding ground truth images to train the machine learning unit. We also used approximately 1000 sample images along with ground truth images for evaluating deep learning methods. The images were acquired from seedlings that were generously supplied by Roelands Plant Farms Inc., the industrial collaborator in this venture.

4. Results and Discussion

We applied our proposed method to five types of seedlings for finding the correct position of the clipping points. These plants are real seedlings from “Roelands Plant Farms” in the stage of clipping task. We evaluated the accuracy of the results comparing the algorithm-selected points with those of the expert farmers. Each category contained 100 images of one type of seedling. Table 2 shows the mAP for each type of seedling in different lighting conditions.
Figure 8 shows the most suitable stem-stake coupling point suggested by KDE-PHOG method to four different seedlings from different lighting conditions and backgrounds. The output of the our suggested method is three parameters. D that is the distance of the camera and suggested clipping point, and ΔθX and ΔθY which are the yaw and pitch correction angle factors. After rotating the robot in yaw and pitch direction with the angle of −ΔθX and −ΔθY, If the robot moves in this direction as much as D, the center of the clipping device will be in the suggested clipping point in the real coordinate of the robot.
After training the deep learning algorithms, we also calculated the mAP obtained using Faster R-CNN, Retina Net, and YOLOv8 for the same set of images. Table 3 shows the mAP for the selected clipping points for each type of seedling using deep learning approaches and compares them with our proposed algorithm.
In real-time applications such as ours and other scenarios where a quick response is required, latency, or computational time, is a critical factor. To compare the computational time of the proposed method and deep learning approaches, we estimated the average and maximum time of finding a clipping point for 500 images. We used the Ryzen7-7730U processor with 16 GB RAM. All codes were ruining on Jupiter notebook (ipykernel:python 3.12.1, TensorFlow 1.15.7, Keras 3.0.2, scikit-learn 1.3.2). Table 4 shows the computational time required for each approach. Average computational time using KDE-PHOG is 124 ms. It means that the process of finding the clipping point is fast and occurs in real time.
The results in the table show that among deep learning approaches, YOLOv8 has a better performance and it even runs slightly faster than our proposed algorithm. However, considering the mPA of all approaches, it is clear that the proposed KDE-PHOG method has the best overall performance.
As seen, all approaches find the clipping point of the bell pepper more accurately. The leaves of both types of cucumbers are big and access to the stem on top of the cucumber seedlings is difficult. So, the mAP of cucumbers is less than other seedlings in all approaches.
In some cases, the stem or stake, or both, could not be accessed—for example, when hidden behind leaves or when the distance between the stem and stake was too great. In such scenarios, the algorithm was unable to suggest a suitable clipping point. As illustrated in Figure 9, for Samples A and B, the algorithm successfully suggests a suitable clipping point. However, for Sample C, it identifies a suitable but unreachable clipping point due to the excessive distance between the stake and stem. For Sample D, neither the stem nor the stake are accessible. A potential solution in these instances is to capture additional images from various angles, or even from the opposite side of the seedling. If the KDE-PHOG algorithm cannot locate the clipping point, the robotic arm has the ability to rotate the clipping device and the stereo camera around the seedling to acquire a new image and identify the clipping point from a different perspective. This approach mirrors the manual methods used by farmers. By implementing this strategy, the mAP of the method is projected to exceed 89 percent.
We designed, built, and installed a clipping device to work alongside our vision system for autonomous clipping. This device was mounted on a general-purpose robotic arm, specifically a KUKA LWR IV. The clipping mechanism operates by curling a thin wire as it attaches the clip to the plant. The clipping device incorporates an optimized stereo camera (Megapixel/USB5000W02M) to capture images of the plants and sends them to the vision algorithm. When the KDE-PHOG algorithm identifies the most suitable clipping point, the robotic arm locates the clipping device near this point to performs the clipping action. Figure 10 displays examples of seedlings with clips attached at the points identified using the KDE-PHOG algorithm. The film “manuscript-supplementary.mp4” shows the operation of an automatic clipping device with a stereo camera on it that are installed on a KUKA LWR IV robot. KDE-PHOG method recognizes the most suitable clipping point and it sends control signals to the robot to guide the robot to the suggested point. The envisioned final automated clipping system will comprise several specialized robotic arms, each equipped with a similar clipping device.

5. Conclusions

This paper introduced a novel approach in machine vision to determine a specific point on the real seedlings in the propagation facilities using the images of a stereo camera. This point is a point on the seedling stem that is suitable for placing a clamp around the stem and fork to provide more support to the seedling during growth or transportation. The proposed vision algorithm is a part of a new robotic stem–stake coupling system under development, which recognizes the suitable point for putting the clip around the stem and stake and puts the clip in the recognized point automatically using robotic arms. Identifying an appropriate clipping point in real-time is challenging due to the complexities involved in mimicking human cognitive processes and incorporating past experiences into the decision-making process. Our proposed approach combines analytical image processing methods with data-driven learning algorithms, setting it apart from deep learning methods. Evaluation of the algorithm was conducted using real seedling images of peppers, tomatoes, and cucumbers. It was shown that our proposed approach resulted in better mPA compared to other methods and an overall best performance. We concluded that the medium natural lighting intensity yielded the best overall results, while bell pepper exhibited the highest mAP, although finding the clipping point on the national pickling cucumber proved more challenging.
To enhance the effectiveness of our system, we implemented a strategy to capture additional images from different angles to reach an mAP of more than 89 percent. This approach emulates the human approach in finding the clipping point. Expert farmers verified the identified clipping points, serving as a valuable ground truth for evaluation.
While the algorithm demonstrates satisfactory performance, there are still some limitations that will be addressed in future studies. Stem recognition can be challenging when leaves cover a significant portion of the stem or when the stem’s pattern and color resemble those of the leaves behind it. To overcome this, employing multiple labeled images to train an Artificial Neural Network or using rational laws such as the continuity of the stem can be explored. Additionally, obtaining multiple images from different angles around the seedlings can be effective to deal with significant occlusions. In our current study, we focused on XGBoost to address non-linear relationships between input features and target variables. we recognize the potential of other deep-learning-based approaches to improve the quality of the results, which will be the focus of our future studies. As a real-time process, optimizing the algorithm’s speed is imperative. Further optimization of the codes for faster throughput can be achieved, and the use of a GPU or optimized hardware will be explored to increase the speed of the process. As we progress towards autonomous seedling clipping, our ongoing research aims to address the limitations, explore learning-based approaches, and optimize the algorithm for enhanced real-time performance.

Supplementary Materials

The following supporting film information can be downloaded at: https://www.mdpi.com/article/10.3390/act13030081/s1, https://uwoca-my.sharepoint.com (accessed on 17 February 2024).

Author Contributions

Conceptualization, M.A.S. and M.R.K.; methodology, M.A.S. and M.R.K.; experiments, M.A.S.; software, M.A.S.; data analysis, M.A.S. and M.R.K.; writing—original draft preparation, M.A.S.; writing—review and editing, M.A.S. and M.R.K.; supervision, M.R.K.; project administration, M.R.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to acknowledge the support received from Adrian Roelands, co-founder of Roelands Plant Farms Inc., and Chris Martin for his assistance and sharing his expertise, which greatly motivated the research and its successful results.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
KDEKernel Density Estimator
PDVPoint Density Variation
PHOGPyramid Histogram of Oriented Gradients
LHGLong Histogram Gradients
PDFProbability Density Function
HOGHistogram of Oriented Gradients
KDE-PHOGReal-time Point Recognition using Kernel Density Estimators and Pyramid Histogram of oriented Gradients
XGBoostExtreme Gradient Boosting
mAPMean Average Precision

References

  1. Xie, D.; Chen, L.; Liu, L.; Chen, L.; Wang, H. Actuators and sensors for application in agricultural robots: A review. Machines 2022, 10, 913. [Google Scholar] [CrossRef]
  2. Berenstein, R.; Edan, Y. Automatic adjustable spraying device for site-specific agricultural application. IEEE Trans. Autom. Sci. Eng. 2017, 15, 641–650. [Google Scholar] [CrossRef]
  3. Wang, Q.; Zhang, Q.; Zhang, Y.; Zhou, G.; Li, Z.; Chen, L. Lodged sugarcane/crop dividers interaction: Analysis of robotic sugarcane harvester in agriculture via a rigid-flexible coupled simulation method. Actuators 2022, 11, 23. [Google Scholar] [CrossRef]
  4. Botta, A.; Cavallone, P.; Baglieri, L.; Colucci, G.; Tagliavini, L.; Quaglia, G. A review of robots, perception, and tasks in precision agriculture. Appl. Mech. 2022, 3, 830–854. [Google Scholar] [CrossRef]
  5. Shafi, U.; Mumtaz, R.; García-Nieto, J.; Hassan, S.A.; Zaidi, S.A.; Iqbal, N. Precision agriculture techniques and practices: From considerations to applications. Sensors 2019, 19, 3796. [Google Scholar] [CrossRef]
  6. Kang, H.; Zhou, H.; Wang, X.; Chen, C. Real-time fruit recognition and grasping estimation for robotic apple harvesting. Sensors 2020, 20, 5670. [Google Scholar] [CrossRef] [PubMed]
  7. Shah, T.M.; Nasika, D.P.; Otterpohl, R. Plant and weed identifier robot as an agroecological tool using artificial neural networks for image identification. Agriculture 2021, 11, 222. [Google Scholar] [CrossRef]
  8. Apostolidis, K.D.; Kalampokas, T.; Pachidis, T.P.; Kaburlasos, V.G. Grapevine Plant Image Dataset for Pruning. Data 2022, 7, 110. [Google Scholar] [CrossRef]
  9. Fan, P.; Lang, G.; Guo, P.; Liu, Z.; Yang, F.; Yan, B.; Lei, X. Multi-feature patch-based segmentation technique in the gray-centered rgb color space for improved apple target recognition. Agriculture 2021, 11, 273. [Google Scholar] [CrossRef]
  10. Zhang, M.; Li, S.; Yu, F.; Tian, X. Image fusion employing adaptive spectral-spatial gradient sparse regularization in uav remote sensing. Signal Process. 2020, 170, 107434. [Google Scholar] [CrossRef]
  11. Shantkumari, M.; Uma, S. Grape leaf segmentation for disease identifi- cation through adaptive snake algorithm model. Multimed. Tools Appl. 2021, 80, 8861–8879. [Google Scholar] [CrossRef]
  12. Shirzi, M.A.; Hairi-Yazdi, M. Active tracking using intelligent fuzzy controller and kernel-based algorithm. In Proceedings of the 2011 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE 2011), Taipei, Taiwan, 27–30 June 2011; pp. 1157–1163. [Google Scholar]
  13. Xiao, D.; Feng, J.; Lin, T.; Pang, C.; Ye, Y. Classification and recognition scheme for vegetable pests based on the BOF-SVM model. Int. J. Agric. Biol. Eng. 2018, 11, 190–196. [Google Scholar] [CrossRef]
  14. Amin, H.; Darwish, A.; Hassanien, A.E.; Soliman, M. End-to-end deep learning model for corn leaf disease classification. IEEE Access 2022, 10, 31103–31115. [Google Scholar] [CrossRef]
  15. Zhang, Q.; Liu, Y.; Gong, C.; Chen, Y.; Yu, H. Applications of deep learning for dense scenes analysis in agriculture: A review. Sensors 2020, 20, 1520. [Google Scholar] [CrossRef] [PubMed]
  16. Kolar, Z.; Chen, H.; Luo, X. Transfer learning and deep convolutional neural networks for safety guardrail detection in 2D images. Autom. Constr. 2018, 89, 58–70. [Google Scholar] [CrossRef]
  17. Benos, L.; Tagarakis, A.C.; Dolias, G.; Berruto, R.; Kateris, D.; Bochtis, D. Machine learning in agriculture: A comprehensive updated review. Sensors 2021, 21, 3758. [Google Scholar] [CrossRef] [PubMed]
  18. Attri, I.; Awasthi, L.K.; Sharma, T.P.; Rathee, P. A review of deep learning techniques used in agriculture. Ecol. Inform. 2023, 77, 102217. [Google Scholar] [CrossRef]
  19. Chen, H.; Heidari, A.A.; Chen, H.; Wang, M.; Pan, Z.; Gandomi, A.H. Multi-population differential evolution-assisted Harris hawks optimization: Framework and case studies. Future Gener. Comput. Syst. 2020, 111, 175–198. [Google Scholar] [CrossRef]
  20. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  21. Das, S.; Gulati, T.; Mittal, V. Histogram equalization techniques for contrast enhancement: A review. Int. J. Comput. Appl. 2015, 114, 32–36. [Google Scholar] [CrossRef]
  22. Thapar, S.; Garg, S. Study and implementation of various morphology based image contrast enhancement techniques. Int. J. Comput. Bus. Res. 2012, 128, 2229–6166. [Google Scholar]
  23. Shirzi, M.A.; Kermani, M. Adaptive Feature-Based Plant Recognition. IEEE Trans. AgriFood Electron. 2023; in press. [Google Scholar]
  24. Ruchay, A.; Kober, V. Impulsive noise removal from color images with morphological filtering. In Proceedings of the Analysis of Images, Social Networks and Texts: 6th International Conference, AIST 2017, Moscow, Russia, 27–29 July 2017; Revised Selected Papers 6 2018 (pp. 280–291). Springer International Publishing: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  25. Kim, B.H.; Cho, J.H. A Study on Modular Smart Plant Factory Using Morphological Image Processing. Electronics 2020, 9, 1661. [Google Scholar] [CrossRef]
  26. Jassim, F.A.; Altaani, F.H. Hybridization of Otsu method and median filter for color image segmentation. arXiv 2013, arXiv:1305.1052. [Google Scholar]
  27. Kumari, D.; Kaur, K. A survey on stereo matching techniques for 3D vision in image processing. Int. J. Eng. Manuf. 2016, 4, 40–49. [Google Scholar] [CrossRef]
  28. Lawin, F.J.; Danelljan, M.; Khan, F.S.; Forssén, P.E.; Felsberg, M. Density adaptive point set registration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3829–3837. [Google Scholar]
  29. Ruckebusch, C. Resolving Spectral Mixtures: With Applications from Ultrafast Time-Resolved Spectroscopy to Super-Resolution Imaging; Elsevier: Amsterdam, The Netherlands, 2016. [Google Scholar]
  30. Yang, C.; Duraiswami, R.; Gumerov, N.A.; Davis, L. Improved fast gauss transform and efficient kernel density estimation. In Proceedings of the Ninth IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 664–671. [Google Scholar]
  31. Elgammal, A.; Duraiswami, R.; Harwood, D.; Davis, L.S. Background and foreground modeling using nonparametric kernel density estimation for visual surveillance. Proc. IEEE 2002, 90, 1151–1163. [Google Scholar] [CrossRef]
  32. Matioli, L.C.; Santos, S.R.; Kleina, M.; Leite, E.A. A new algorithm for clustering based on kernel density estimation. J. Appl. Stat. 2018, 45, 347–366. [Google Scholar] [CrossRef]
  33. Scaldelai, D.; Matioli, L.C.; Santos, S.R.; Kleina, M. MulticlusterKDE: A new algorithm for clustering based on multivariate kernel density estimation. J. Appl. Stat. 2022, 49, 98–121. [Google Scholar] [CrossRef]
  34. Chen, Y.-C.; Genovese, C.R.; Wasserman, L. A comprehensive approach to mode clustering. Electron. J. Stat. 2016, 10, 210–241. [Google Scholar] [CrossRef]
  35. Gour, N.; Khanna, P. Automated glaucoma detection using GIST and pyramid histogram of oriented gradients (PHOG) descriptors. Pattern Recognit. Lett. 2020, 137, 3–11. [Google Scholar] [CrossRef]
  36. Xie, L.; Lee, F.; Liu, L.; Yin, Z.; Yan, Y.; Wang, W.; Zhao, J.; Chen, Q. Improved spatial pyramid matching for scene recognition. Pattern Recognit. 2018, 82, 118–129. [Google Scholar] [CrossRef]
  37. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13 August 2016; pp. 785–794. [Google Scholar]
  38. Elavarasan, D.; Vincent, D.R. Reinforced XGBoost machine learning model for sustainable intelligent agrarian applications. J. Intell. Fuzzy Syst. 2020, 39, 7605–7620. [Google Scholar] [CrossRef]
  39. Yoon, H.I.; Lee, H.; Yang, J.S.; Choi, J.H.; Jung, D.H.; Park, Y.J.; Park, J.E.; Kim, S.M.; Park, S.H. Predicting Models for Plant Metabolites Based on PLSR, AdaBoost, XGBoost, and LightGBM Algorithms Using Hyperspectral Imaging of Brassica juncea. Agriculture 2023, 13, 1477. [Google Scholar] [CrossRef]
  40. Ding, X.; Li, Q.; Cheng, Y.; Wang, J.; Bian, W.; Jie, B. Local keypoint-based Faster R-CNN. Appl. Intell. 2020, 50, 3007–3022. [Google Scholar] [CrossRef]
  41. Bao, W.; Fan, T.; Hu, G.; Liang, D.; Li, H. Detection and identification of tea leaf diseases based on AX-RetinaNet. Sci. Rep. 2022, 12, 2183. [Google Scholar] [CrossRef]
  42. Fang, W.; Wang, L.; Ren, P. Tinier-YOLO: A real-time object detection method for constrained environments. IEEE Access 2019, 8, 1935–1944. [Google Scholar] [CrossRef]
  43. Li, X.; Zeng, L.; Zheng, L. Improvement of the key point detection algorithm based on yolov8. In Proceedings of the International Conference on Algorithms, High Performance Computing, and Artificial Intelligence (AHPCAI 2023), Yinchuan, China, 7 December 2023; Volume 12941, pp. 206–213. [Google Scholar]
Figure 1. Manual clipping of seedlings. The images show the density of the seedlings and their arrangement in a typical propagation facility.
Figure 1. Manual clipping of seedlings. The images show the density of the seedlings and their arrangement in a typical propagation facility.
Actuators 13 00081 g001
Figure 2. Experts’ selections of clipping points on cucumber, tomato, and pepper seedlings. The yellow circle with the red arrow show the suggested clipping points from expert farmers.
Figure 2. Experts’ selections of clipping points on cucumber, tomato, and pepper seedlings. The yellow circle with the red arrow show the suggested clipping points from expert farmers.
Actuators 13 00081 g002
Figure 3. Block diagram of the real-time seedling point recognition algorithm using kernel density estimators and pyramid histogram of oriented gradients.
Figure 3. Block diagram of the real-time seedling point recognition algorithm using kernel density estimators and pyramid histogram of oriented gradients.
Actuators 13 00081 g003
Figure 4. Steps for computing the orientation of the histogram gradient. The red arrow shows the size and angle of the POHG.
Figure 4. Steps for computing the orientation of the histogram gradient. The red arrow shows the size and angle of the POHG.
Actuators 13 00081 g004
Figure 5. Segmented plant and its skeleton and boundary. The red and blue lines shows the boundary of the ROI.
Figure 5. Segmented plant and its skeleton and boundary. The red and blue lines shows the boundary of the ROI.
Actuators 13 00081 g005
Figure 6. Steps of selcting candidate cell using XGBoost. For example, histogram of PDV, feature vector of KDE and normalized PHOG of four areas of the left images including areas A, B, C and D are given in the right table.
Figure 6. Steps of selcting candidate cell using XGBoost. For example, histogram of PDV, feature vector of KDE and normalized PHOG of four areas of the left images including areas A, B, C and D are given in the right table.
Actuators 13 00081 g006
Figure 7. Steps for finding clipping point after finding the candidate cell. The red and blue lines show the position of the suitable clipping point on images. Green colors show the position of stem and stake in the y-direction in the suggested clipping point.
Figure 7. Steps for finding clipping point after finding the candidate cell. The red and blue lines show the position of the suitable clipping point on images. Green colors show the position of stem and stake in the y-direction in the suggested clipping point.
Actuators 13 00081 g007
Figure 8. Suggested clipping points after applying KDE-PHOG method to different types of seedlings from different distances, and with different lighting conditions and backgrounds, three parameters are calculated related to the most suitable clipping point which are D, ΔθX and ΔθY which are the distance of the camera and suggested clipping point, the yaw and pitch correction angle factors.
Figure 8. Suggested clipping points after applying KDE-PHOG method to different types of seedlings from different distances, and with different lighting conditions and backgrounds, three parameters are calculated related to the most suitable clipping point which are D, ΔθX and ΔθY which are the distance of the camera and suggested clipping point, the yaw and pitch correction angle factors.
Actuators 13 00081 g008
Figure 9. Suggested clipping points using multi-stage point density algorithm for samples (A,B). In sample (C), a clipping point is identified but the stake and stem are too far apart. Sample (D) shows a case where neither the stem nor the stake is accessible.
Figure 9. Suggested clipping points using multi-stage point density algorithm for samples (A,B). In sample (C), a clipping point is identified but the stake and stem are too far apart. Sample (D) shows a case where neither the stem nor the stake is accessible.
Actuators 13 00081 g009
Figure 10. The stereo camera takes images from the seedling and stake to identify a suitable clipping point using KDE-PHOG method. The robotic arm moves the clipping device near the recognized clipping point, and the clipping device makes a clip around the stem and stake. The yellow circles indicates the position of the clips that have been made and put on the suitable clipping point using automatic clipping device to couple the stem of the seedlings and stake.
Figure 10. The stereo camera takes images from the seedling and stake to identify a suitable clipping point using KDE-PHOG method. The robotic arm moves the clipping device near the recognized clipping point, and the clipping device makes a clip around the stem and stake. The yellow circles indicates the position of the clips that have been made and put on the suitable clipping point using automatic clipping device to couple the stem of the seedlings and stake.
Actuators 13 00081 g010
Table 1. A brief description of methods: Faster R-CNN, Faster R-CNN and YOLOv8.
Table 1. A brief description of methods: Faster R-CNN, Faster R-CNN and YOLOv8.
NameDescription
Faster R-CNNFaster R-CNN has set new standards in object detection by seamlessly integrating the region proposal mechanism into the overall architecture. It provides a balance of high accuracy and efficiency, establishing itself as a fundamental model in the field of computer vision for tasks that demand accurate object localization and classification. It is primarily designed for instance segmentation, which involves identifying and delineating individual objects in an image [40].
RetinaNetRetinaNet is an effective one-stage object detection model that employs focal loss to handle the imbalance between positive and negative samples in training. Its primary focus lies in predicting both bounding boxes and class probabilities for objects present in an image [41]. RetinaNet is adapted to predict the coordinates of points by including specific output layers.
YOLOv8As a form of instance segmentation, YOLO combines image processing and deep learning techniques to achieve real-time object detection in both images and videos [42]. YOLOv8 can be adapted for point recognition with some modifications and by adding an extension [43].
Table 2. Mean Average Precision (mAP) of selected clipping points using KDE-PHOG method for each seedling in different lighting conditions.
Table 2. Mean Average Precision (mAP) of selected clipping points using KDE-PHOG method for each seedling in different lighting conditions.
Natural Light IntensityArtificial
SeedlingLowMediumHighLight (LED)Average
Tomato83%90%87%88%87%
Bell Pepper91%96%94%95%94%
Chili Pepper88%94%89%93%91%
National Pickling Cucumber81%88%84%87%85%
Beit Alpha Cucumber82%91%86%87%87%
Standard Deviation3.487
Total Average89%
Table 3. Comparison of mean Average Precision (mAP) of KDE-PHOG and deep learning approaches in selecting the clipping point.
Table 3. Comparison of mean Average Precision (mAP) of KDE-PHOG and deep learning approaches in selecting the clipping point.
SeedlingFaster R-CNNRetinaNetYOLOv8KDE-PHOG
Tomato75%78%85%87%
Bell Pepper80%83%90%94%
Chili Pepper81%85%82%91%
National Pickling Cucumber74%82%84%85%
Beit Alpha Cucumber71%78%80%87%
Standard Deviation3.7632.7863.3713.249
Average76%81%84%89%
Table 4. Average and maximum computational time for a complete cycle of finding the clipping point.
Table 4. Average and maximum computational time for a complete cycle of finding the clipping point.
Computational Time (ms)Faster R-CNNRetinaNetYOLOv8KDE-PHOG
Average16314592124
Maximum208164116156
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Asadi Shirzi, M.; Kermani, M.R. Real-Time Point Recognition for Seedlings Using Kernel Density Estimators and Pyramid Histogram of Oriented Gradients. Actuators 2024, 13, 81. https://doi.org/10.3390/act13030081

AMA Style

Asadi Shirzi M, Kermani MR. Real-Time Point Recognition for Seedlings Using Kernel Density Estimators and Pyramid Histogram of Oriented Gradients. Actuators. 2024; 13(3):81. https://doi.org/10.3390/act13030081

Chicago/Turabian Style

Asadi Shirzi, Moteaal, and Mehrdad R. Kermani. 2024. "Real-Time Point Recognition for Seedlings Using Kernel Density Estimators and Pyramid Histogram of Oriented Gradients" Actuators 13, no. 3: 81. https://doi.org/10.3390/act13030081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop