Next Article in Journal
Phenolic Profiles and Bioactivities of Ten Original Lineage Beans in Thailand
Next Article in Special Issue
Development and Performance Analysis of an Automatic Core Cutter for Elephant Apple (Dillenia indica L.) Processing
Previous Article in Journal
Investigating the Probiotic Properties and Antimicrobial Activity of Lactic Acid Bacteria Isolated from an Iranian Fermented Dairy Product, Kashk
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Coconut Clusters Based on Occlusion Condition Using Attention-Guided Faster R-CNN for Robotic Harvesting

by
L. G. Divyanth
1,
Peeyush Soni
1,*,
Chaitanya Madhaw Pareek
1,
Rajendra Machavaram
1,
Mohammad Nadimi
2 and
Jitendra Paliwal
2,*
1
Department of Agricultural and Food Engineering, Indian Institute of Technology Kharagpur, Kharagpur 721302, India
2
Department of Biosystems Engineering, University of Manitoba, Winnipeg, MB R3T 5V6, Canada
*
Authors to whom correspondence should be addressed.
Foods 2022, 11(23), 3903; https://doi.org/10.3390/foods11233903
Submission received: 22 September 2022 / Revised: 23 November 2022 / Accepted: 1 December 2022 / Published: 3 December 2022
(This article belongs to the Special Issue Digital Innovation in Agricultural and Food Technology)

Abstract

:
Manual harvesting of coconuts is a highly risky and skill-demanding operation, and the population of people involved in coconut tree climbing has been steadily decreasing. Hence, with the evolution of tree-climbing robots and robotic end-effectors, the development of autonomous coconut harvesters with the help of machine vision technologies is of great interest to farmers. However, coconuts are very hard and experience high occlusions on the tree. Hence, accurate detection of coconut clusters based on their occlusion condition is necessary to plan the motion of the robotic end-effector. This study proposes a deep learning-based object detection Faster Regional-Convolutional Neural Network (Faster R-CNN) model to detect coconut clusters as non-occluded and leaf-occluded bunches. To improve identification accuracy, an attention mechanism was introduced into the Faster R-CNN model. The image dataset was acquired from a commercial coconut plantation during daylight under natural lighting conditions using a handheld digital single-lens reflex camera. The proposed model was trained, validated, and tested on 900 manually acquired and augmented images of tree crowns under different illumination conditions, backgrounds, and coconut varieties. On the test dataset, the overall mean average precision (mAP) and weighted mean intersection over union (wmIoU) attained by the model were 0.886 and 0.827, respectively, with average precision for detecting non-occluded and leaf-occluded coconut clusters as 0.912 and 0.883, respectively. The encouraging results provide the base to develop a complete vision system to determine the harvesting strategy and locate the cutting position on the coconut cluster.

1. Introduction

Coconut (Cocos nucifera) is one of the main cash crops grown throughout the year, predominantly along the coastal regions of Asia, South America, and the Pacific Islands. Coconut cultivation globally accounts for about 11.8 million hectares of agricultural land across 93 countries and nearly 63 million tonnes per year [1]. Among the Asian countries that contribute to almost 75% of the global production (52 million tonnes), Indonesia (17.1 million tonnes), Philippines (14.8 million tonnes), and India (14.7 million tonnes) account for more than 90% of the coconut production [2].
Coconut products have become an indispensable part of our day-to-day lives and for numerous industries in many parts of the world due to their high nutritional value and wide range of food and industrial products [1]. However, farmers are often trapped in financial losses despite increased demand for coconut products due to costly human labour and arduous work in many stages of its production. Manual harvesting of coconuts is one such operation that involves many operational risks and safety issues because of the tree’s height and hard trunk. Conventionally, coconuts are harvested by specially trained and experienced climbers who climb to the tree top and then cut the coconut bunches. This operation is hazardous, as falling from a tree due to slippage can result in severe irrecoverable physical damage and sometimes death [3]. Because of this high-risk manual harvesting operation, most young people are not interested in adopting it as an occupation despite decent wages. This scenario has led to an acute shortage of trained and skilled labour for this high-skill-demanding operation, especially during peak harvesting season.
In order to mechanize this operation, various tree climbing devices have been proposed [4,5,6,7]. Abraham et al. [8] developed a remote-controlled coconut tree climbing robot in which the ground operator can regulate the locomotion and harvesting operations with the help of a camera mounted on a robot. As an improvement, Dubey et al. [9] proposed an autonomous climbing robot; however, its harvesting and spraying mechanisms needed to be manually controlled.
With the rapid advances in artificial intelligence and robotics technologies over the past decade, autonomous vision-based robotic harvesting has emerged as a preferred approach for the selective harvesting of various fruits and vegetables [10,11]. In general, a typical vision-based fruit-harvesting robot includes two main subsystems: a machine-vision system and an end-effector system. The vision system detects and localizes the harvesting targets, and then the end-effector approaches these targets. The successful execution of the harvesting task requires accurate detection and localization of the harvesting targets [10,12]. Moreover, information about the robotic arm’s working environment is crucial for obstacle-free access to these targets.
Thus, the effectiveness of robot end-effector manipulation depends greatly on the performance of the fruit detection algorithm. With the recent advancements in computational resources and the evolution of deep learning as a promising tool, researchers are increasingly relying on deep learning algorithms for developing vision-based harvesting systems [13,14,15]. Unlike traditional computer vision solutions that relied on identifying important features and tedious feature-engineering tasks [16,17,18], deep learning uses convolutional neural networks (CNNs) to automatically learn important features within training data, resulting in a less biased model. Deep learning algorithms have shown potential in many machine-vision operations applied to agriculture [19,20,21,22]. Numerous studies have employed CNN-based fruit detection models to support robotic harvesting, such as in the case of apples [23,24], tomatoes [25], strawberries [26], mangoes [27], kiwifruit [28], and blueberries [29].
In the case of the robotic harvesting of coconuts, the robotic end-effector could be damaged if the robot attempts to harvest occluded coconut clusters, which results in poor efficiency and economic loss. In recently developed coconut harvesting machines, the end-effector consists of a sharp circular high-speed rotating blade intended to cut the stalk of coconut bunches. For instance, Megalingam et al. [7] proposed a robotic arm with four degrees of freedom and a rotatory cutter end-effector. However, such an end-effector is very likely to damage the tree crown. Therefore, a crucial step for an automated harvesting system is to approach the coconut bunches via a path that avoids collision with the occlusions. For that purpose, the coconut bunches should be detected according to the occlusion condition. Thus, the objective of this study was to develop a deep learning object-detection model based on an improved Faster R-CNN algorithm to detect the coconut clusters as two classes: non-occluded coconuts and leaf-occluded coconuts.

2. Related Work

The performance of the fruit detection algorithm greatly determines the efficiency of autonomous harvesting. Over the past few years, there has been remarkable progress in developing smart fruit detection systems [13]. The conventional methods of fruit detection involve deriving the colour, texture, and morphology-based features present in the images [30,31]. However, they suffer from poor detection performance and robustness and fall short of the working needs of robotic harvesters. These image-processing-based algorithms are also prone to errors in unstructured backgrounds, occlusions, and different varieties of fruit (such as in the case of coconuts). In recent years, CNN-based algorithms have shown great advantages in fruit detection [32,33]. Many researchers have studied multi-class fruit identification in a variety of orchards for robotic harvesting [32]. Among various deep learning models, the Faster R-CNN family of networks has shown great potential for detecting objects of different sizes and aspect ratios, which makes them extensively useful for fruit detection [23,24,28,34,35,36].
In this regard, Fu et al. [23] used Faster R-CNN for detecting apples from RGB-D images of dense foliage orchards. They compared the performances of ZFNet and VGG16 networks for feature extraction, and the highest mean precision value of 0.89 was achieved by VGG16. The same networks as the feature extraction component of Faster R-CNN were compared for their result of multi-class apple fruit detection based on the occlusion condition [24]. Again, the VGG16 achieved the highest mean precision of 0.88 for the four-class detection at an average speed of 0.24 s per image. In another study, the Faster R-CNN model achieved the highest F1-score of 0.84 for sweet pepper detection [34]. Wan et al. proposed an improved Faster R-CNN network for the detection of apples, mangoes, and oranges [36]. Their model showcased a mean precision value of 0.86, which was higher than that of YOLOv3. Parvathi and Selvi [37] proposed a Faster R-CNN-based model to detect two main coconut maturity stages of consumer interest, namely, ripened and tender coconuts, under complex field conditions. The authors reported that the performance was better than other object-detection models, such as YOLOv3, R-FCN, and Single Shot Detector (SSD). Henceforth, motivated by results from these studies, we adopt the Faster R-CNN network for developing the machine-vision system for detecting coconuts based on their occlusion conditions.
Moreover, some previous studies have implemented an attention module to further optimize the deep learning model. The attention mechanism focuses more on the important features and reduces the huge number of intermediate features. For instance, a small CNN model (with 2.13 million parameters) with an attention mechanism and feature fusion module was designed to identify wheat ear diseases in field conditions [38]. The mentioned study achieved an accuracy of 94.1% and reduced the impact of complex backgrounds. The mechanism has also shown potential for the segmentation [39] and detection tasks [40]. The recent YOLOv4-tiny model was enhanced with adaptive feature fusion and attention modules to detect green peppers for developing a robotic picking system [41]. Notably, this model improved occluded and overlapped green pepper detection with an average precision of 95.1%, which was better than the performance of the algorithm tested without the attention mechanism. Hence, an attention module was additionally attached to the Faster R-CNN model to highlight and captivate the necessary features for coconuts detection.

3. Materials and Methods

3.1. Dataset and Software

3.1.1. Image Acquisition

The image data for this study were collected during a harvest season (July 2021) on a commercial coconut plantation in Western Tamil Nadu, India (Figure 1). The height of the coconut trees on the farm ranged from around 8 m to 20 m. The temperature and humidity ranged between 25–35 °C and 78–83%, respectively, with occasional clouds during the daytime. A total of 900 Red-Green-Blue (RGB) images of coconut tree crowns were acquired during daylight between 08:00 to 17:00 (morning to evening) under natural lighting conditions using a handheld digital single-lens reflex (DSLR) camera (Nikon-D3200, 40–60 mm lens, with a resolution of 300 × 300 dpi and 6016 × 4000 pixels). The experimenter stood on the ground near the tree trunk and used the camera to photograph the tree crown. Since the camera of the vision system in a coconut tree climbing robot would be facing towards the crown, the images collected in this study (the experimenter also points the camera upwards towards the tree crown) were analogous to the ones that would be captured by the system. Images were acquired standing at multiple locations around the tree so that different sections of the tree crown were covered. The path aside from every plantation row was used by the experimenter to move to subsequent trees.
The dataset experienced variation in the illumination conditions when the sun angle changed, which could aid in the development of a more robust detection model. Because the ultimate motive of this study was to develop a machine vision platform for an automated coconut harvester (real-time application), the images were collected at different positions (elevation), angles, and lighting conditions (frontal, back, and scattered lighting) to ensure effective generalization of the model. The images also featured diverse conditions associated with the background, shadows, and tree crown orientation. The images were resized to 512 × 512 pixels since the original size was too large and demanded high computational resources and time for detection. A few images from the dataset are presented in Figure 2. The dataset can be accessed at https://data.mendeley.com/datasets/w4t73tvrf8.
A Windows 10-based system with MATLAB® R2021b (Mathworks Inc., Natick, MA, USA) software was used for the analysis. The Image Processing Toolbox, Statistics and Machine Learning Toolbox, and Deep Learning Toolbox were additionally installed into the MATLAB® software.

3.1.2. Coconut Classes and Ground Truth Preparation

Unlike apples and berries, coconuts cannot be harvested through common robotic approaches like picking or shaking mechanisms due to their hard nature. The region of impact on the coconut cluster for robotic harvesting is often occluded by the petiole, cushion, leaflets, or other parts of the tree crown (Figure 3). Hence, the coconut clusters were categorized into two classes based on occlusion (Figure 4). The first ‘non-occluded coconuts’ class represents the independent coconut clusters that can be harvested directly with priority. The second ‘leaf-occluded coconuts’ class refers to the coconut clusters occluded by the various portions of the coconut leaf.
The coconut clusters were annotated into two classes with rectangular bounding boxes (Figure 5). All images were manually annotated using the interactive MATLAB® Image Labeler app. The following abbreviations/class names were used to indicate the classes (which will be followed in the rest of the article): ‘NOC’ for non-occluded coconuts, and ‘LOC’ for leaf-occluded coconuts.

3.1.3. Image Augmentation

To increase the size of the dataset and prevent the model from overfitting, geometry- and intensity-based image augmentations were additionally performed. Image geometry-based augmentation techniques that were adopted include random rotations at 90°, 180°, and 270° as well as horizontal and vertical flipping. While performing the above geometry transformations, a small code strip was written in the MATLAB® environment to automatically calibrate the bounding box coordinates of the transformed images from the original image’s bounding box annotations. Furthermore, image intensity transformations included brightness variation with propositional coefficients of 0.8, 0.9, 1.1, and 1.2 as well as histogram equalization after converting the RGB images to their HSV (Hue-Saturation-Value) equivalents. The blurring of images is a potential problem in vision-based harvesting systems due to the camera’s movement. To make the detection network more adaptable to such real conditions, a Gaussian filter with a standard deviation value of two was applied to the images as another augmentation strategy for artificial blurring.
Thus, the size of the dataset was increased by 12 times (from 900 to 10,800 sample images) through data augmentation, which was randomly separated as follows: 6800 images for training the model; 2000 images for validation; and 2000 images for testing purposes. The class-wise distribution of coconut clusters in the dataset is presented in Table 1.

3.2. Deep Learning Network Architecture

The architecture of the proposed coconut detection model is detailed in this section. The structure of the improved Faster R-CNN model is illustrated in Figure 6, which consists of a features extraction network, a Region Proposal Network (RPN), and a final classification network. To improve the quality of features derived from the feature extraction network, an attention mechanism was affixed to this feature extraction network to concentrate on the informational channels more than the less important channels.

3.2.1. Feature Extraction Network and Attention Module

In the proposed model, the VGG16 network was incorporated as the backbone feature extraction network. The network comprises four maximum-pooling layers that divide the thirteen convolutional layers into five groups. These pooling layers help extract a broad range of feature maps from different levels of the network. To enhance the feature representation, this was followed by an attention module (before feeding the feature maps into the RPN) that amended and improved the feature maps and provoked a better quality of region proposals by emphasizing the descriptive feature channels.
In the attention module, a global average pooling operation is initially performed on the input feature map to convert them into a channel descriptor. The descriptor is then passed into two successive fully connected layers, which were respectively modulated by the ReLU and sigmoid function. This arrangement aids in gaining the properties of a channel from a global perspective and handling channel-wise interdependencies. The output from the second fully connected layer (modulated by sigmoid function) accords an attention descriptor β , whose values emulate the importance of the respective channels. If X H × W × C denotes the input features to the attention module, the H × W dimensional feature maps were compressed to a channel descriptor z 1 × C by the dense layer as in Equation (1).
z c = 1 H × W   i = 1 H j = 1 W X c ( i , j )  
The descriptor z is converted into the probabilistic attention descriptor ( β 1 × 1 × C ) by the sigmoid operation. Thus, the input feature channel maps are recalibrated in accordance with the attention descriptor by multiplying this channel-by-channel (Equation (2)) to generate informative feature maps with more emphasis.
Y C n = F h a d a m a r d ( X , β ) = β X
where F h a d a m a r d denotes the Hadamard product (channel-wise element-wise multiplication). The attention module used in the coconut detector is presented in Figure 7. The final feature maps are fed into the RPN for region proposal.

3.2.2. Region Proposal Network and Classification Network

A typical region-based object detection CNN is based on the following steps: (i) identify the region that might contain the object (called the region proposals), (ii) extract the features from the proposed regions, and (iii) classify the objects and adjust the bounding box. The RPN, which is also a fully convolutional network, is used to create the region proposals using the features from the VGG16 feature extractor. The region of interest pooling operation (ROI-pooling) is carried out to transform these region proposals (which do not have a uniform size) into the same size. In the ROI-pooling layer, the non-uniform region proposals are divided into a fixed dimension, say N w × N h sub-blocks, and a max-pooling operation is carried out in each sub-block. This constructs a feature map of fixed sizes with the most salient features. These fixed-size bounding box guesses are called anchors, which are uniformly placed throughout the image to capture the aspect ratio of target objects. The typical dimensions (height and width) of anchor boxes ( N w × N h ) were chosen based on the size of the coconut clusters in the images.
The classification network leverages the RPN’s proposals and the feature maps from the attention module as the inputs. This network consisted of two convolution layers and two fully connected layers followed by two branches (a region classification branch and a regression branch) for fine-tuning the anchor box coordinates. Based on the scores and predictions of the classification layer, the positions of the plausible anchor boxes can be traced back from the output of the CNN back to the input image. Once the position of an anchor box is mapped, the localization errors (between the ground truth bounding box and the predicted tiled anchor box) are fixed. Through the regression branch, the detector learns the offsets to be applied to the anchor box to adjust its position and size.

3.3. Network Training

The attention-guided Faster R-CNN network with VGG16 as the feature extraction network was trained for the task of coconut cluster detection. The fully convolutional RPN was optimized using the back-propagation algorithm and mini-batch gradient descent, with a loss function comprising the sum of the classification and bounding box regression losses. The hyperparameters set during the training process are provided in Table 2. The MATLAB® application was run on Acer Nitro 5 Intel Core i5 9th Generation Laptop (32 GB/1 TB HDD/Windows 10 Home/GTX 1650 Graphics).

3.4. Performance Evaluation Metrics

The performance of the detection model was evaluated using the weighted mean Intersection-over-Union (wmIoU) and mean Average Precision (mAP) metrics.

3.4.1. Weighted Mean Intersection-over-Union

Intersection-over-Union (IoU), or the Jaccard similarity coefficient or Jaccard index, is a measure of the overlap between the ground truth and predicted bounding boxes. The IoU can be determined via the ratio of the area of overlap to the area of the union of two bounding boxes (Figure 8). Predicted bounding boxes that have heavily overlapped with the ground truth have a higher IoU value compared to ones with less overlap. If A p r e d i c t e d denotes the predicted bounding box and A g r o u n d t r u t h is the ground truth, then IoU is mathematically expressed as Equation (3).
I o U = A p r e d i c t e d     A g r o u n d t r u t h   A p r e d i c t e d     A g r o u n d t r u t h
Intuitively, the mean IoU (mIoU) for an image is the average of IoU values of the predicted bounding box with the ground truth for each class in the image. In this work, to reduce the biasing effect of IoU for dominating classes, the mIoU was weighted with the number of objects in each class ( X c ) for all images in the dataset, called the weighted mean IoU (wmIoU). The wmIoU can be expressed in terms of mIoU for N images and C classes as in Equation (4).
w m I o U = 1 N × C × c = 1 C X c 1 n = 1 N m I o U

3.4.2. Mean Average Precision

In this study, a detection was considered a positive prediction if the IoU was greater than 0.5. With that, all detections were classified as true positive (TP), false positive (FP), true negative (TN), or false negative (FN). Thus, precision (P) and recall (R) were defined as:
P r e c i s i o n   ( P ) = T P T P + F P
R e c a l l   ( R ) = T P T P + F N
The global performance of a network to detect the target object is determined by its Average Precision (AP), defined as the area under the precision–recall curve (Equation (7)), which is obtained by plotting P and R in the vertical and horizontal axes, respectively. If APc represents the AP of a particular class c, then the mean AP (mAP) is given as the average precision of all classes (Equation (8)). Hence, the greater the value of APc, the better the results of the coconut detection system. The value of c represents the classes considered in the experiment: NOC (c = 1) and LOC (c = 2).
A P c = 0 1 P ( R c )   d R c
m A P = 1 C c = 1 C A P c

4. Results and Discussion

In order to verify the performance of the attention-guided Faster R-CNN model for the detection of coconut clusters, 2000 images were used to test the model. The mAP and wmIoU values for the test data were 0.88 and 0.83, respectively. The class-wise detection results on test and validation sets are presented in Table 3, and example images are illustrated in Figure 9. Since NOCs are completely visible, they were easier to be detected with a higher mIoU (0.90) and AP rate (0.91). The LOCs were also identified with better precision, but the mIoU value (0.80) was relatively low. This shows that although the LOC clusters were detected with high precision, the regression loss due to localization of the bounding boxes was high, thus resulting in reduced IoU measure (which indicates the proximity of predicted and ground truth bounding boxes). The model was also able to effectively detect coconut clusters occluded by many leaves (Figure 9).
The model took 0.774 s on average to detect coconut clusters in images of size 512 × 512 pixels, which was deemed adequate for developing a real-time vision-based harvester. The size of the model’s trained weights was 548 MB, comparable to other Faster R-CNN-based models in the literature designed for real-time fruit detection [13,24]. Because the coconut detection system would potentially be deployed in a tree climbing robot, the size and number of clusters in the field of view will differ with its movement on the tree trunk. The size would be larger, and fewer coconuts would be visible when the vision system is near the tree crown, hence making it easier to detect. However, more clusters will be present in the field of view when the robot is farther. Therefore, the model’s performance was analyzed for test images that were acquired closer to the tree crown and farther away from it (Figure 10). The images were considered to be taken farther from the crown when more than approximately 20% of the length of the tree was captured in the image. Inspection of the images showed that bounding boxes detected for close images had a better prediction of the ground truth than the farther images that had intensive coconut clusters. Nevertheless, the result for closer images is more significant since harvesting is performed at the time when the robot is near the tree crown. Moreover, it is to be noted that the anchor box size used by the model can have a deliberate impact on the recognition of coconuts of different sizes and the box’s aspect ratio.
The results imply that the model can facilitate the detection of different numbers of coconut clusters from the image captured from any position on the tree trunk. Including an attention mechanism provided higher weight to the features of coconuts, thus improving the ability to recognize coconuts with a range of occlusions. As shown in Figure 11, the proposed method was suitable for images with different illumination conditions. In addition, the detection results were impressive for images collected under direct sunlight (Figure 11a,b) and backlight (Figure 11c) conditions. The method was also applicable for the detection of poorly illuminated coconut clusters under strongly backlit conditions (Figure 11d). However, the manual analysis showed that the IoU value was lower for detection in images with intensive coconut clusters under backlighting. Unlike the detection of small fruit such as apples and strawberries, which experienced poor performance due to shadows and similarities with backgrounds [26,42], coconut detection under such conditions appears relatively easier due to its large size and unique shape among the objects in the image’s field of view.
Practically, the colour of coconuts differs greatly in a plantation due to the accommodation of many varieties, which comprise shades of green, yellow, orange, and brown. It was observed through the experiment that this did not have any impact on the prediction performance, meaning that the principal features utilized for coconut detection were shape- and texture-based. Additionally, very accurate detection of the coconut clusters when their edges are occluded is problematic since their sizes and shape vary greatly (unlike fruit such as apples, whose shape can be very closely determined even when partially occluded). In real-time, the vision system may also capture images without any coconuts during the robot’s operation. To test the validity of the model in this case, ten tree crown images that did not contain any coconut were derived from internet sources. As expected, no predictions were reported by the model in these images.
Since the features of coconuts belonging to different clusters in the same tree are very similar, identifying and differentiating the coconuts of two bunches when one occluded another was challenging (Figure 12). Although considerable detection of the coconut cluster causing the occlusion was possible, the model faced difficulty in placing the prediction box coordinates of the occluded coconut cluster.
Overall, this study statistically analyzed the performance of a deep learning approach for coconut detection to assist the robot in planning the harvesting strategy based on occlusion conditions. The proposed model proved effective in detecting the coconut clusters into two classes with a high precision rate (Table 3). Many previous studies have stressed the potential of the attention mechanism in Faster R-CNN for improving the result as well as reducing the prediction time [43,44,45]. In the case of leaf-occlusion, either a special path must be adapted for the end-effector to reach the coconut bunch, or the leaf must be cut down before proceeding to harvest in order to free the cluster from obstruction. In that case, if many leaves occlude a cluster, more research should be carried out in the future to locate the position of leaves as well for efficient harvesting. With good results obtained through the attention-guided Faster R-CNN model for the detection of a complete coconut bunch, further investigation on model development is intended to be carried out for localizing and estimating the position of a cutting point in the cluster.
Multi-class detection based on occlusion conditions for harvesting has been widely studied for other fruits, including apples [24,42], tomatoes [46], cereal grains [47,48], and citrus fruits [49]. Similar to the proposed model, Gao et al. [24] used Faster R-CNN with VGG-16 as a feature extraction network for detecting apples and classifying them into four classes based on occlusion (non-, leaf-, branch/wire-, and fruit-occluded). The model achieved mAP of 0.88 and AP of 0.91, 0.90, 0.86, and 0.85 for the respective classes. In a similar study, Mu et al. [46] used R-CNN with Resnet-101 to detect occluded tomatoes and achieved an mAP of 0.88. There have been no previously reported studies on real-time coconut detection for harvesting or other robotic applications. Hence, it is difficult to compare the results of this study with others. To the best of our knowledge, only Parvathi and Selvi [37] have worked on the development of a deep learning-based coconut detection system for maturity stage identification. They tested four algorithms, namely, Faster R-CNN, SSD, YOLOv3, and FCN, and concluded that the results achieved by Faster R-CNN (mAP = 0.89) were significantly better than the other models. Importantly, the maturity identification approach can be combined with occlusion condition analysis to give a better perception of the vision system for decisive coconut harvesting. However, such a model demands more classes to be comprehended. Another interesting topic for future research is the performance of similar state-of-the-art deep learning models in robotic systems for other applications such as grading agri-food products [50,51,52,53,54,55,56,57,58], monitoring crop diseases [59,60,61], and assessing crop growth and yield [62,63,64,65,66].

5. Conclusions

In this study, a deep learning model based on state-of-the-art Faster R-CNN with VGG-16 was proposed for detecting coconuts based on their occlusion condition. The analysis showed that the model achieved high precision and speed for identifying non-occluded and leaf-occluded coconuts, which can evade potential damage to the tree and robot end-effector during robotic harvesting. The attention mechanism was introduced into the Faster R-CNN model to improve the ability of the network to identify occluded coconuts. The model achieved mAP of 0.88 and wmIoU of 0.83 for coconut detection in the two classes, and the speed was 0.77 s per image. The proposed method offers valuable information in planning the end-effector movement for coconut harvesting safely and efficiently.

Author Contributions

Conceptualization, L.G.D. and P.S; methodology, L.G.D. and P.S; software, L.G.D.; validation, L.G.D., P.S., C.M.P. and R.M.; formal analysis, L.G.D.; investigation, L.G.D., P.S. and R.M.; resources, L.G.D., P.S., M.N. and J.P.; data curation, L.G.D.; writing—original draft preparation, L.G.D. and C.M.P.; writing—review and editing, P.S., C.M.P., R.M., M.N. and J.P.; supervision, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Council of Canada (NSERC) grant number RGPIN2021-03350. The authors would also like to thank the financial support provided by Mitacs.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ignacio, I.F.; Miguel, T.S. Research Opportunities on the Coconut (Cocos Nucifera L.) Using New Technologies. South Afr. J. Bot. 2021, 141, 414–420. [Google Scholar] [CrossRef]
  2. Kabir Ahmad, R.; Anwar Sulaiman, S.; Yusup, S.; Sham Dol, S.; Inayat, M.; Aminu Umar, H. Exploring the Potential of Coconut Shell Biomass for Charcoal Production. Ain Shams Eng. J. 2022, 13, 101499. [Google Scholar] [CrossRef]
  3. Maciej, S.; Becker, F.G.; Cleary, M.; Team, R.M.; Holtermann, H.; The, D.; Agenda, N.; Science, P.; Sk, S.K.; Hinnebusch, R.; et al. Health of Coconut Tree Climbers of Rural Southern India—Medical Emergencies, Body Mass Index and Occupational Marks: A Quantitative and Survey Study. J. Clin. Diagnostic Res. 2012, 6, 57–60. [Google Scholar] [CrossRef]
  4. Author, C.; Purushottamrao Kolhe, K.; Powar, A.G.; Dhakane, A.D.; Mankar, S.H. Stability and Ergonomic Design Features of Tractor Mounted Hydraulic Elevator for Coconut Harvesting. Am. J. Eng. Appl. Sci. 2011, 4, 380–389. [Google Scholar] [CrossRef]
  5. Lam, T.L.; Xu, Y. Climbing Strategy for a Flexible Tree Climbing Robot—Treebot. IEEE Trans. Robot. 2011, 27, 1107–1117. [Google Scholar] [CrossRef]
  6. Kuttankulangara Manoharan, S.; Megalingam, R.K. Coconut Tree Structure Analysis—Background Work for an Unmanned Coconut Harvesting Robot Design. Commun. Comput. Inf. Sci. 2019, 835, 207–222. [Google Scholar] [CrossRef]
  7. Megalingam, R.K.; Manoharan, S.K.; Mohandas, S.M.; Vadivel, S.R.R.; Gangireddy, R.; Ghanta, S.; Kumar, K.S.; Teja, P.S.; Sivanantham, V. Amaran:An Unmanned Robotic Coconut Tree Climber and Harvester. IEEE/ASME Trans. Mechatron. 2021, 26, 288–299. [Google Scholar] [CrossRef]
  8. Abraham, A.; Girish, M.; Vitala, H.R.; Praveen, M.P. Design of Harvesting Mechanism for Advanced Remote-Controlled Coconut Harvesting Robot (A.R.C.H-1). Indian J. Sci. Technol. 2014, 7, 1465–1470. [Google Scholar] [CrossRef]
  9. Dubey, A.P.; Pattnaik, S.M.; Banerjee, A.; Sarkar, R.; Kumar, S.R. Autonomous Control and Implementation of Coconut Tree Climbing and Harvesting Robot. Procedia Comput. Sci. 2016, 85, 755–766. [Google Scholar] [CrossRef] [Green Version]
  10. Zhao, Y.; Gong, L.; Huang, Y.; Liu, C. A Review of Key Techniques of Vision-Based Control for Harvesting Robot. Comput. Electron. Agric. 2016, 127, 311–323. [Google Scholar] [CrossRef]
  11. Montoya-Cavero, L.E.; Díaz de León Torres, R.; Gómez-Espinosa, A.; Escobedo Cabello, J.A. Vision Systems for Harvesting Robots: Produce Detection and Localization. Comput. Electron. Agric. 2022, 192, 106562. [Google Scholar] [CrossRef]
  12. Vasconez, J.P.; Kantor, G.A.; Auat Cheein, F.A. Human–Robot Interaction in Agriculture: A Survey and Current Challenges. Biosyst. Eng. 2019, 179, 35–48. [Google Scholar] [CrossRef]
  13. Koirala, A.; Walsh, K.B.; Wang, Z.; McCarthy, C. Deep Learning—Method Overview and Review of Use for Fruit Detection and Yield Estimation. Comput. Electron. Agric. 2019, 162, 219–234. [Google Scholar] [CrossRef]
  14. Hasan, A.S.M.M.; Sohel, F.; Diepeveen, D.; Laga, H.; Jones, M.G.K. A Survey of Deep Learning Techniques for Weed Detection from Images. Comput. Electron. Agric. 2021, 184, 106067. [Google Scholar] [CrossRef]
  15. Osco, L.P.; Marcato Junior, J.; Marques Ramos, A.P.; de Castro Jorge, L.A.; Fatholahi, S.N.; de Andrade Silva, J.; Matsubara, E.T.; Pistori, H.; Gonçalves, W.N.; Li, J. A Review on Deep Learning in UAV Remote Sensing. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102456. [Google Scholar] [CrossRef]
  16. Nguyen, T.T.; Vandevoorde, K.; Wouters, N.; Kayacan, E.; De Baerdemaeker, J.G.; Saeys, W. Detection of Red and Bicoloured Apples on Tree with an RGB-D Camera. Biosyst. Eng. 2016, 146, 33–44. [Google Scholar] [CrossRef]
  17. Liu, X.; Zhao, D.; Jia, W.; Ji, W.; Sun, Y. A Detection Method for Apple Fruits Based on Color and Shape Features. IEEE Access 2019, 7, 67923–67933. [Google Scholar] [CrossRef]
  18. Singh, N.; Tewari, V.K.; Biswas, P.K.; Pareek, C.M.; Dhruw, L.K. Image Processing Algorithms for In-Field Cotton Boll Detection in Natural Lighting Conditions. Artif. Intell. Agric. 2021, 5, 142–156. [Google Scholar] [CrossRef]
  19. Divyanth, L.G.; Ahmad, A.; Saraswat, D. A Two-Stage Deep-Learning Based Segmentation Model for Crop Disease Quantification Based on Corn Field Imagery. Smart Agric. Technol. 2022, 3, 100108. [Google Scholar] [CrossRef]
  20. Divyanth, L.G.; Marzougui, A.; González-Bernal, M.J.; McGee, R.J.; Rubiales, D.; Sankaran, S. Evaluation of Effective Class-Balancing Techniques for CNN-Based Assessment of Aphanomyces Root Rot Resistance in Pea (Pisum Sativum L.). Sensors 2022, 22, 7237. [Google Scholar] [CrossRef]
  21. Moazzam, S.I.; Khan, U.S.; Tiwana, M.I.; Iqbal, J.; Qureshi, W.S.; Shah, S.I. A Review of Application of Deep Learning for Weeds and Crops Classification in Agriculture. In Proceedings of the 2019 International Conference on Robotics and Automation in Industry (ICRAI), Rawalpindi, Pakistan, 21–22 October 2019. [Google Scholar] [CrossRef]
  22. Divyanth, L.G.; Chelladurai, V.; Loganathan, M.; Jayas, D.S.; Soni, P. Identification of Green Gram (Vigna Radiata) Grains Infested by Callosobruchus Maculatus Through X-ray Imaging and GAN-Based Image Augmentation. J. Biosyst. Eng. 2022, 47, 302–317. [Google Scholar] [CrossRef]
  23. Fu, L.; Majeed, Y.; Zhang, X.; Karkee, M.; Zhang, Q. Faster R–CNN–Based Apple Detection in Dense-Foliage Fruiting-Wall Trees Using RGB and Depth Features for Robotic Harvesting. Biosyst. Eng. 2020, 197, 245–256. [Google Scholar] [CrossRef]
  24. Gao, F.; Fu, L.; Zhang, X.; Majeed, Y.; Li, R.; Karkee, M.; Zhang, Q. Multi-Class Fruit-on-Plant Detection for Apple in SNAP System Using Faster R-CNN. Comput. Electron. Agric. 2020, 176, 105634. [Google Scholar] [CrossRef]
  25. Zhang, L.; Jia, J.; Gui, G.; Hao, X.; Gao, W.; Wang, M. Deep Learning Based Improved Classification System for Designing Tomato Harvesting Robot. IEEE Access 2018, 6, 67940–67950. [Google Scholar] [CrossRef]
  26. Lamb, N.; Chuah, M.C. A Strawberry Detection System Using Convolutional Neural Networks. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2019; pp. 2515–2520. [Google Scholar] [CrossRef]
  27. Ranjan, A.; Machavaram, R. Detection and Localisation of Farm Mangoes Using YOLOv5 Deep Learning Technique. In Proceedings of the 2022 IEEE 7th International Conference for Convergence in Technology (I2CT), Mumbai, India, 7–9 April 2022. [Google Scholar] [CrossRef]
  28. Fu, L.; Feng, Y.; Majeed, Y.; Zhang, X.; Zhang, J.; Karkee, M.; Zhang, Q. Kiwifruit Detection in Field Images Using Faster R-CNN with ZFNet. IFAC-PapersOnLine 2018, 51, 45–50. [Google Scholar] [CrossRef]
  29. Ni, X.; Li, C.; Jiang, H.; Takeda, F. Three-Dimensional Photogrammetry with Deep Learning Instance Segmentation to Extract Berry Fruit Harvestability Traits. ISPRS J. Photogramm. Remote Sens. 2021, 171, 297–309. [Google Scholar] [CrossRef]
  30. Zhao, C.; Lee, W.S.; He, D. Immature Green Citrus Detection Based on Colour Feature and Sum of Absolute Transformed Difference (SATD) Using Colour Images in the Citrus Grove. Comput. Electron. Agric. 2016, 124, 243–253. [Google Scholar] [CrossRef]
  31. Kapach, K.; Barnea, E.; Mairon, R.; Edan, Y.; Ben-Shahar, O. Computer Vision for Fruit Harvesting Robots—State of the Art and Challenges Ahead. Int. J. Comput. Vis. Robot. 2012, 3, 4–34. [Google Scholar] [CrossRef] [Green Version]
  32. Ukwuoma, C.C.; Zhiguang, Q.; Bin Heyat, M.B.; Ali, L.; Almaspoor, Z.; Monday, H.N. Recent Advancements in Fruit Detection and Classification Using Deep Learning Techniques. Math. Probl. Eng. 2022, 2022, 9210947. [Google Scholar] [CrossRef]
  33. Gongal, A.; Amatya, S.; Karkee, M.; Zhang, Q.; Lewis, K. Sensors and Systems for Fruit Detection and Localization: A Review. Comput. Electron. Agric. 2015, 116, 8–19. [Google Scholar] [CrossRef]
  34. Sa, I.; Ge, Z.; Dayoub, F.; Upcroft, B.; Perez, T.; McCool, C. DeepFruits: A Fruit Detection System Using Deep Neural Networks. Sensors 2016, 16, 1222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Gené-Mola, J.; Vilaplana, V.; Rosell-Polo, J.R.; Morros, J.R.; Ruiz-Hidalgo, J.; Gregorio, E. Multi-Modal Deep Learning for Fuji Apple Detection Using RGB-D Cameras and Their Radiometric Capabilities. Comput. Electron. Agric. 2019, 162, 689–698. [Google Scholar] [CrossRef]
  36. Wan, S.; Goudos, S. Faster R-CNN for Multi-Class Fruit Detection Using a Robotic Vision System. Comput. Networks 2020, 168, 107036. [Google Scholar] [CrossRef]
  37. Parvathi, S.; Tamil Selvi, S. Detection of Maturity Stages of Coconuts in Complex Background Using Faster R-CNN Model. Biosyst. Eng. 2021, 202, 119–132. [Google Scholar] [CrossRef]
  38. Bao, W.; Yang, X.; Liang, D.; Hu, G.; Yang, X. Lightweight Convolutional Neural Network Model for Field Wheat Ear Disease Identification. Comput. Electron. Agric. 2021, 189, 106367. [Google Scholar] [CrossRef]
  39. Kang, J.; Liu, L.; Zhang, F.; Shen, C.; Wang, N.; Shao, L. Semantic Segmentation Model of Cotton Roots In-Situ Image Based on Attention Mechanism. Comput. Electron. Agric. 2021, 189, 106370. [Google Scholar] [CrossRef]
  40. Dong, S.; Wang, R.; Liu, K.; Jiao, L.; Li, R.; Du, J.; Teng, Y.; Wang, F. CRA-Net: A Channel Recalibration Feature Pyramid Network for Detecting Small Pests. Comput. Electron. Agric. 2021, 191, 106518. [Google Scholar] [CrossRef]
  41. Li, X.; Pan, J.; Xie, F.; Zeng, J.; Li, Q.; Huang, X.; Liu, D.; Wang, X. Fast and Accurate Green Pepper Detection in Complex Backgrounds via an Improved Yolov4-Tiny Model. Comput. Electron. Agric. 2021, 191, 106503. [Google Scholar] [CrossRef]
  42. Wang, D.; He, D. Channel Pruned YOLO V5s-Based Deep Learning Approach for Rapid and Accurate Apple Fruitlet Detection before Fruit Thinning. Biosyst. Eng. 2021, 210, 271–281. [Google Scholar] [CrossRef]
  43. Zeng, W.; Li, M. Crop Leaf Disease Recognition Based on Self-Attention Convolutional Neural Network. Comput. Electron. Agric. 2020, 172, 105341. [Google Scholar] [CrossRef]
  44. Gao, R.; Wang, R.; Feng, L.; Li, Q.; Wu, H. Dual-Branch, Efficient, Channel Attention-Based Crop Disease Identification. Comput. Electron. Agric. 2021, 190, 106410. [Google Scholar] [CrossRef]
  45. Xia, X.; Chai, X.; Zhang, N.; Sun, T. Visual Classification of Apple Bud-Types via Attention-Guided Data Enrichment Network. Comput. Electron. Agric. 2021, 191, 106504. [Google Scholar] [CrossRef]
  46. Mu, Y.; Chen, T.S.; Ninomiya, S.; Guo, W. Intact Detection of Highly Occluded Immature Tomatoes on Plants Using Deep Learning Techniques. Sensors 2020, 20, 2984. [Google Scholar] [CrossRef] [PubMed]
  47. Wang, W.; Paliwal, J. Spectral Data Compression and Analyses Techniques to Discriminate Wheat Classes. Trans. ASABE 2006, 49, 1607–1612. [Google Scholar] [CrossRef]
  48. Mebatsion, H.K.; Paliwal, J. A Fourier Analysis Based Algorithm to Separate Touching Kernels in Digital Images. Biosyst. Eng. 2011, 108, 66–74. [Google Scholar] [CrossRef]
  49. Lu, J.; Sang, N. Detecting Citrus Fruits and Occlusion Recovery under Natural Illumination Conditions. Comput. Electron. Agric. 2015, 110, 121–130. [Google Scholar] [CrossRef]
  50. Natarajan, S.; Ponnusamy, V. Agri-Food Products Quality Assessment Methods. In Computer Vision and Machine Learning in Agriculture; Algorithms for Intelligent Systems; Uddin, M.S., Bansal, J.C., Eds.; Springer: Singapore, 2022; Volume 2. [Google Scholar] [CrossRef]
  51. Hosainpour, A.; Kheiralipour, K.; Nadimi, M.; Paliwal, J. Quality assessment of dried white mulberry (Morus alba L.) using machine vision. Horticulturae 2022, 8, 1011. [Google Scholar] [CrossRef]
  52. Li, X.; Guillermic, R.M.; Nadimi, M.; Paliwal, J.; Koksel, F. Physical and microstructural quality of extruded snacks made from blends of barley and green lentil flours. Cereal Chem. 2022, 99, 1112–1123. [Google Scholar] [CrossRef]
  53. Sivakumar, C.; Chaudhry, M.M.A.; Nadimi, M.; Paliwal, J.; Courcelles, J. Characterization of Roller and Ferkar-Milled Pulse Flours Using Laser Diffraction and Scanning Electron Microscopy. Powder Technol. 2022, 409, 117803. [Google Scholar] [CrossRef]
  54. Sabzi, S.; Abbaspour, Y.; Nadimi, M.; Paliwal, J. Non-destructive estimation of physicochemical properties and detection of ripeness level of apples using machine vision. Int. J. Fruit Sci. 2022, 22, 628–645. [Google Scholar] [CrossRef]
  55. Divyanth, L.G.; Guru, D.S.; Soni, P.; Machavaram, R.; Nadimi, M.; Paliwal, J. Image-to-Image Translation-Based Data Augmentation for Improving Crop/Weed Classification Models for Precision Agriculture Applications. Algorithms 2022, 15, 401. [Google Scholar] [CrossRef]
  56. Nadimi, M.; Loewen, G.; Paliwal, J. Assessment of Mechanical Damage to Flaxseeds Using Radiographic Imaging and To-mography. Smart Agric. Technol. 2022, 2, 100057. [Google Scholar] [CrossRef]
  57. Nadimi, M.; Divyanth, L.G.; Paliwal, J. Automated detection of mechanical damage in flaxseeds using radiographic imaging and machine learning. Foods Bioprocess Technol. 2022. [Google Scholar] [CrossRef]
  58. Kheiralipour, K.; Nadimi, M.; Paliwal, J. Development of an Intelligent Imaging System for Ripeness Determination of Wild Pistachios. Sensors 2022, 22, 7134. [Google Scholar] [CrossRef] [PubMed]
  59. Ouhami, M.; Hafiane, A.; Es-Saady, Y.; El Hajji, M.; Canals, R. Computer Vision, IoT and Data Fusion for Crop Disease Detection Using Machine Learning: A Survey and Ongoing Research. Remote Sens. 2021, 13, 2486. [Google Scholar] [CrossRef]
  60. Nadimi, M.; Brown, J.M.; Morrison, J.; Paliwal, J. Examination of wheat kernels for the presence of Fusarium damage and mycotoxins using near-infrared hyperspectral imaging. Meas. Food 2021, 4, 100011. [Google Scholar] [CrossRef]
  61. Erkinbaev, C.; Nadimi, M.; Paliwal, J. A Unified Heuristic Approach to Simultaneously Detect Fusarium and Ergot Damage in Wheat. Meas. Food 2022, 7, 100043. [Google Scholar] [CrossRef]
  62. Ruangurai, P.; Dailey, M.N.; Ekpanyapong, M.; Soni, P. Optimal Vision-Based Guidance Row Locating for Autonomous Agricultural Machines. Precis. Agric. 2022, 23, 1205–1225. [Google Scholar] [CrossRef]
  63. Sirikun, C.; Samseemoung, G.; Soni, P.; Langkapin, J.; Srinonchat, J. A Grain Yield Sensor for Yield Mapping with Local Rice Combine Harvester. Agric. 2021, 11, 897. [Google Scholar] [CrossRef]
  64. Nadimi, M.; Loewen, G.; Bhowmik, P.; Paliwal, J. Effect of laser biostimulation on germination of sub-optimally stored flaxseeds (Linum usitatissimum). Sustainability 2022, 14, 12183. [Google Scholar] [CrossRef]
  65. Nadimi, M.; Sun, D.W.; Paliwal, J. Effect of laser biostimulation on germination of wheat. ASABE Appl. Eng. Agric. 2022, 38, 77–84. [Google Scholar] [CrossRef]
  66. Nadimi, M.; Sun, D.W.; Paliwal, J. Recent applications of novel laser techniques for enhancing agricultural production. Laser Phys. 2021, 31, 053001. [Google Scholar] [CrossRef]
Figure 1. Commercial coconut plantation from where the dataset was collected.
Figure 1. Commercial coconut plantation from where the dataset was collected.
Foods 11 03903 g001
Figure 2. Sample images in the coconut dataset.
Figure 2. Sample images in the coconut dataset.
Foods 11 03903 g002
Figure 3. Occluded coconuts and structure of coconut leaf.
Figure 3. Occluded coconuts and structure of coconut leaf.
Foods 11 03903 g003
Figure 4. Coconuts categorized into two classes based on the condition of occlusion: (a) non-occluded coconuts and (b) leaf-occluded coconuts.
Figure 4. Coconuts categorized into two classes based on the condition of occlusion: (a) non-occluded coconuts and (b) leaf-occluded coconuts.
Foods 11 03903 g004
Figure 5. Rectangular annotations of coconut targets for creating ground truth. The blue and yellow bounding boxes refer to the non-occluded (NOC) and leaf-occluded (LOC) coconut targets, respectively.
Figure 5. Rectangular annotations of coconut targets for creating ground truth. The blue and yellow bounding boxes refer to the non-occluded (NOC) and leaf-occluded (LOC) coconut targets, respectively.
Foods 11 03903 g005
Figure 6. Illustration of the architecture of the proposed coconut cluster detection system based on Faster R-CNN with VGG16 and the attention module.
Figure 6. Illustration of the architecture of the proposed coconut cluster detection system based on Faster R-CNN with VGG16 and the attention module.
Foods 11 03903 g006
Figure 7. Illustration of the attention module employed in the proposed detection model.
Figure 7. Illustration of the attention module employed in the proposed detection model.
Foods 11 03903 g007
Figure 8. Illustration of the intersection over union (IoU) evaluation metric.
Figure 8. Illustration of the intersection over union (IoU) evaluation metric.
Foods 11 03903 g008
Figure 9. Detection of coconut clusters using the proposed model. The labels show the detected class (NOC: Non-Occluded Coconuts; LOC: Leaf-Occluded Coconuts) and confidence score.
Figure 9. Detection of coconut clusters using the proposed model. The labels show the detected class (NOC: Non-Occluded Coconuts; LOC: Leaf-Occluded Coconuts) and confidence score.
Foods 11 03903 g009
Figure 10. Coconut detection result concerning tree crown proximity: (a) near to the tree crown and (b) farther from the tree crown.
Figure 10. Coconut detection result concerning tree crown proximity: (a) near to the tree crown and (b) farther from the tree crown.
Foods 11 03903 g010
Figure 11. Detection of coconut clusters under different illumination: (a,b) direct sunlight, and (c,d) backlight conditions.
Figure 11. Detection of coconut clusters under different illumination: (a,b) direct sunlight, and (c,d) backlight conditions.
Foods 11 03903 g011
Figure 12. Misdetection of coconut-occluded coconut clusters; two clusters were detected as a single cluster of coconuts.
Figure 12. Misdetection of coconut-occluded coconut clusters; two clusters were detected as a single cluster of coconuts.
Foods 11 03903 g012
Table 1. Class-wise distribution of coconut clusters in the training, validation, and test datasets after performing image augmentation.
Table 1. Class-wise distribution of coconut clusters in the training, validation, and test datasets after performing image augmentation.
ClassNumber of Coconut ClustersTotal Number of Coconut Clusters in the Class (in Complete Dataset)
TrainingValidationTest
Non-occluded coconut clusters76742220231012,204
Leaf-occluded coconut clusters11,2502856281416,920
Total 18,9245076512429,124
Table 2. Some hyperparameters set during training of the model.
Table 2. Some hyperparameters set during training of the model.
HyperparameterValue
OptimizerStochastic gradient descent with momentum (sgdm)
Initial learn rate0.001
Maximum epochs1000
Mini batch size32
Learn rate drop factor0.0005
Learn rate drop period10
Momentum0.9
Table 3. Class-wise results for coconut cluster detection using attention-guided Faster R-CNN.
Table 3. Class-wise results for coconut cluster detection using attention-guided Faster R-CNN.
mIoUAverage Precision (AP)
ValidationTestValidationTest
Non-occluded coconuts0.9060.8950.9240.912
Leaf-occluded coconuts0.8120.8070.8990.883
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Divyanth, L.G.; Soni, P.; Pareek, C.M.; Machavaram, R.; Nadimi, M.; Paliwal, J. Detection of Coconut Clusters Based on Occlusion Condition Using Attention-Guided Faster R-CNN for Robotic Harvesting. Foods 2022, 11, 3903. https://doi.org/10.3390/foods11233903

AMA Style

Divyanth LG, Soni P, Pareek CM, Machavaram R, Nadimi M, Paliwal J. Detection of Coconut Clusters Based on Occlusion Condition Using Attention-Guided Faster R-CNN for Robotic Harvesting. Foods. 2022; 11(23):3903. https://doi.org/10.3390/foods11233903

Chicago/Turabian Style

Divyanth, L. G., Peeyush Soni, Chaitanya Madhaw Pareek, Rajendra Machavaram, Mohammad Nadimi, and Jitendra Paliwal. 2022. "Detection of Coconut Clusters Based on Occlusion Condition Using Attention-Guided Faster R-CNN for Robotic Harvesting" Foods 11, no. 23: 3903. https://doi.org/10.3390/foods11233903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop