Next Article in Journal
The Spatiotemporal Decoupling Relationship between Economic Development, Energy Consumption, and Carbon Dioxide Emissions in Xinjiang Province from 2006 to 2020
Previous Article in Journal
The US Economy as a Network: A Comparison across Economic and Environmental Metrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Growth Stages of Chilli Plants in a Hydroponic Grower Using Machine Vision and YOLOv8 Deep Learning Algorithms

by
Florian Schneider
,
Jonas Swiatek
and
Mohieddine Jelali
*
Cologne Laboratory of Artificial Intelligence and Smart Automation (CAISA), Institute of Product Development and Engineering Design (IPK), Technische Hochschule Köln—University of Applied Sciences, 50679 Cologne, Germany
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(15), 6420; https://doi.org/10.3390/su16156420
Submission received: 17 April 2024 / Revised: 21 June 2024 / Accepted: 24 July 2024 / Published: 26 July 2024

Abstract

:
Vertical indoor farming (VIF) with hydroponics offers a promising perspective for sustainable food production. Intelligent control of VIF system components plays a key role in reducing operating costs and increasing crop yields. Modern machine vision (MV) systems use deep learning (DL) in combination with camera systems for various tasks in agriculture, such as disease and nutrient deficiency detection, and flower and fruit identification and classification for pollination and harvesting. This study presents the applicability of MV technology with DL modelling to detect the growth stages of chilli plants using YOLOv8 networks. The influence of different bird’s-eye view and side view datasets and different YOLOv8 architectures was analysed. To generate the image data for training and testing the YOLO models, chilli plants were grown in a hydroponic environment and imaged throughout their life cycle using four camera systems. The growth stages were divided into growing, flowering, and fruiting classes. All the trained YOLOv8 models showed reliable identification of growth stages with high accuracy. The results indicate that models trained with data from both views show better generalisation. YOLO’s middle architecture achieved the best performance.

1. Introduction

Thanks to vertical indoor farming (VIF) with hydroponics, salads, fruit, vegetables, and spices could grow directly in the city in the future. The pioneers of the technology [1] hope that vertical indoor farms, also known as vertical plant factories, can ensure a greater food supply for the rapidly growing world population and will lead to greater sustainability and better flavour. At present, VIF is still a niche market, but it is recognised as an expanding sector.
To date, most vertical plant factories have focused primarily on the production of leafy vegetables, salads, and micro-vegetables, but also many spices/herbs, as their production cycles and environmental requirements are relatively simple. In contrast, there is only a very limited selection of fruit and fruiting vegetables that can be grown well in a vertically controlled environment. Fruit plants such as tomatoes, peppers, cucumbers, and strawberries, among others, are much more difficult to cultivate than other plants. They need more nutrients, more light, and more care than leafy vegetables or herbs [2].
Chilli peppers are used in countless dishes in many cultures and are appreciated for their flavour and spiciness. There are around 4000 varieties of chilli worldwide. These are divided into five varieties: Capsicum annuum (C. annuum), Capsicum baccatum, Capsicum chinense, Capsicum frutescens, and Capsicum pubescens. The largest and industrially most important variety is C. annuum [3], which is considered in this study. Chillies were harvested as early as 8000 BC and were only native to the American continent until the discovery of America in 1492 [4]. The pods of chilli plants are a rich source of vitamins and contain antioxidants, which are of therapeutic importance in the treatment of metabolic disorders and obesity [5]. Species of C. annuum can grow up to 1.5 m tall and their flowers and fruits usually hang downwards. The growth temperature is 25 degrees Celsius [6].
Chilli plants belong to the nightshade family and are divided into 10 macro-stages according to the BBCH scale (Biologische Bundesanstalt für Land- und Forstwirtschaft, Bundessortenamt und Chemische Industrie), which shows the sequence of the main growth stages. This detailed description covers plant growth from germination to death [7]. The scale was updated in 2021 by Feldmann et al. [8]. Paul et al. [9] divided the chilli plants into three growth stages to examine them for fungal infestation. Paul et al. [10] divided the growth cycle of pepper plants into six stages for image recognition: buds, flowers, unripe, ripe, and overripe peppers.
Machine vision systems with convolutional neural networks (CNNs) are used in many areas of modern agriculture, for example, to classify plant varieties, count fruit, detect diseases, localise weeds, automate harvesting, and monitor fields [11,12]. Many reviews have reported the MV application status in agriculture. They are mainly involved in field crops, for example, [13], but only a few of them refer to plant factories, for example, [12].
Automatic monitoring systems are implemented in indoor farms for the visual inspection of plants. The cameras are mounted in a bird’s-eye view and are positioned above the plants using moving axes [14]. For example, the cultivation of seedlings is monitored [15]. These systems are also used to measure the size of plants with stereo cameras and are suitable for vertical cultivation [16,17]. Camera motion systems are also used in the field of plant phenotyping, which quantitatively analyses and measures the external appearance (phenotype) of plants [18,19].
You Only Look Once (YOLO) models, introduced in [20], are mainly used in the field of object recognition for fruit detection and achieve very good results in terms of accuracy, processing speed, and the processing of high-resolution images [21]. Coleman et al. [22] investigated the recognition of cotton growth stages using various YOLO models. The v8x model performed best for recognising the eight classes considered; a precision of up to 89.7% was achieved. Paul et al. [10] have successfully developed a YOLOv8 model for recognising the growth stages of pepper plants in greenhouses. Their dataset consists of images mainly from the side view, focused on flowers and fruits. They state that YOLOv8m and YOLOv8n emerged as superior choices for capsicum detection and growth stage determination, achieving mean average precision (mAP) scores of 95.2% and 75.1%, respectively. The recognition of growth stages in hydroponic systems has been successfully implemented on lettuce using images from above only [23]. The authors achieved 96.0% for recall, precision, and F1 score using a modified YOLOv5 version. Xiao et al. [24] proposed an improved YOLOv5 algorithm for the recognition of blueberry fruit maturity stages. The algorithm achieved a precision of 96.3%, a recall of 92%, and an mAP of 91.5%.
The challenges that MV systems and algorithms face in VIF are [12]:
  • Changing lighting conditions and complex indoor backgrounds make it difficult for MV algorithms to segment feature areas from images. In addition to the plant itself, there are irrigation pipes, suspension cables, mechanical equipment, and other support facilities. The lighting also changes periodically according to the needs of the plants, i.e., growth stages.
  • There are gaps in the knowledge of the application of MV in specific indoor scenarios, which affect the effectiveness of the technology.
In this paper, the recognition of the growth stages from two configurations: (a) a bird’s-eye view and (b) the combined bird’s-eye and side views are presented and discussed. The effects on the accuracy of extending the datasets with images from the side view as well as the performance of different YOLOv8 model architectures are investigated. In contrast, the state of the art is that camera systems are placed only above the planting bed. The growth stages of the chilli plant are divided into three classes: growing, flowering, and fruiting. This categorisation results from the necessary adjustment of the hydroponic system parameters in relation to light and nutrient solution. Furthermore, we install industrial cameras while most references use digital cameras to obtain images manually, which inevitably causes uneven and non-reproducible image quality.
The remaining sections of the paper are organised as follows: Section 2 describes the experimental setup (hydroponic environment developed for cultivation of chilli plants) and introduces the methods for pre-processing the image datasets, including image acquisition, image data augmentation, the creation of image datasets, and the methods and algorithms (YOLOv8) implemented for the detection of the growth stages of chilli plants. Section 3 presents the experimental results and a comparative analysis. In Section 4, the results are discussed. Section 5 provides the conclusions and directions for subsequent work and improvement.

2. Materials and Methods

For the investigation of the growth stages of chilli plants in VIF from a bird’s-eye and side views, Figure 1 shows the contributions and development scope of this work. The process includes the image acquisition, the experimental setup, and the acquisition and storage of images. The next step is the computer vision system, which includes image pre-processing and modelling using DL and evaluation.

2.1. Experimental Field

Figure 2 shows the experimental setup with the chilli plants and the four camera systems installed. The overall hydroponics test stand system developed consists of several subsystems: Essentially, a hydroponics system (including the nutrient system), a lighting system, sensor systems, and a hydroponics kit are integrated in a cultivation box. The experiments presented in this work were carried out in the CAISA Lab at the Cologne University of Applied Sciences (TH Köln, Cologne, Germany). The DiamondBox Silver Line SL150 (azzeo GmbH, Regensburg, Germany) [25] grow box was used for the trials. The dimensions of the system are 1500 mm in width, 1500 mm in depth, and 2000 mm in height.

2.1.1. Hydroponics System and Planting

Plant growth requires a controlled working environment so that the growth parameters can be permanently checked and guaranteed. This includes the monitoring of temperature and humidity in the atmosphere as well as an even air flow. pH and EC measurements were used to control the nutrient solution and the photoperiod could be adjusted.
The hydroponic system used was the GrowStream10 V2 NFT system from Terra Aquatica (Smilinggreen.de, Schlüchtern, Germany), which is particularly suitable for growing chilli plants [2]. It consists of two double-walled plastic troughs, in each of which five plants can be placed. The storage container for the nutrient solution has a capacity of 45 L. The overall dimensions of the system are 1130 mm in length, 460 mm in width, and 530 mm in height. For the light supply, which is essential for photosynthesis, the Lumatek ATTIS ATS 300 W PRO LED (Lumatek Ltd., Rochester, UK) was used, which is designed for a cultivation area of 1 m2. A GHP Profan clip fan was used to ensure the necessary air circulation within the grow box.
The composition of the nutrient solution was based on an N-P-K ratio of 3-1-3 (nitrogen—N–phosphoric anhydride—P2O–potassium oxide—K2O). Both a dry fertiliser Peters Professional Grow-Mix and a liquid fertiliser Terra Aquatica TriPart were used for fertilisation. Terra Aquatica pH-Down and pH-Up were used to regulate the pH value. A disinfectant concentrate was added to the nutrient solution as a prophylactic measure to prevent algae growth. The transplanted areas were sealed with a 3D-printed lid to prevent the penetration of light, which would promote algae growth.
The pH value for growth was kept between 5.8 and 6.5 and the EC value between 1.4 and 1.8 mS/cm. These values, including the temperature of the nutrient solution, were automatically controlled by an Atlas Wi-Fi Hydroponics Kit and manually regulated. The light intensity was measured with a Quantum PAR Light Meter from Lightscout at the top edge of the net baskets, which are located in the plastic gutter. The PPFD (photosynthetically active photon flux density) value at a height of 14 cm was 257 µmol/sm2 and at a height of 58 cm, 584 µmol/s2 (verified by measurements). The photoperiod was set to 9 h until the first flowering and to 12 h for further growth. A combination meter displayed humidity and temperature.
In this study, plants were examined at growing (or seedling), flowering, and fruiting stages of chilli pepper growth, as defined in [9]: growing stage means the early stage of chilli pepper growth, i.e., 10 days of seedling transplanting when 5–7 leaves have appeared; flowering refers to the stage in the growth cycle when the first flowers appear on the plant, i.e., after 25 days of planting; fruiting stage marks the period when the plant enters the fruiting phase, approximately 40 days after planting, and the chilli plants begin to develop fruits.

2.1.2. Camera Systems

In order to train a machine learning model effectively, a suitable dataset must be created. In the field of computer vision, the data consist of images. In order to generate a large number of images, four cameras were placed in the grow box. Two cameras were used for the bird’s-eye view and two cameras for the side view. As shown in the schematic system setup in Figure 3, the distances between the cameras and the chilli plants depend on the respective growth stages. The system was designed to record and store images of at least two chilli plants over their entire life cycle and to consider the changing plants in their entirety, as they can reach a height of between 500 and 700 mm and a width of 300 to 400 mm. To ensure sharp images during growth, flexible positioning of the camera systems is possible.
A plant, with its countless leaves, flowers, and fruits is a complex representation. To train a deep learning model, it needs a detailed representation of the object. The HQ Raspberry Pi camera (Sony Europe B.V., Pencoed, Wales, UK) with a 12 megapixel camera with a 1/4″ (3.2 mm) image sensor was used for this application. The sensor size B, the object width g, and the object size G are of central importance for the design of the lens. These values can be used to calculate the focal length of the lens using the BB formula f = B B + G g . The expected plant height or object size G was 300 mm in the growing phase and 600 mm from the flowering phase onwards. The object width g was set at 500 mm in the first phase and 800 mm from the second phase onwards. The object distance g for the cameras from the bird’s-eye view was set to a maximum of 500 mm. This resulted in focal lengths f of 5.28 mm and 4.24 mm. A 6 mm wide-angle lens is available for the HQ Raspberry Pi camera. To process and save the images, the camera system was connected to the compatible Raspberry Pi Zero W single-board computer (Sony Europe B.V., Pencoed, Wales, UK), including a 32 GB memory card.
The camera system for the side view (camera systems 1 and 3) was attached to a height-adjustable travelling stand using a mounting bracket, as shown in Figure 3a. For the bird’s-eye view camera system (camera systems 2 and 4), a frame made of aluminium profiles was positioned inside the grow box. Two perpendicular profiles were mounted on the cross strut, to which the mounting bracket for the camera system was attached, as illustrated in Figure 3b. Table 1 lists the components required for the camera system, including the Raspberry Pi 4 B for the central image storage and network cable for the power supply.

2.2. Image Data Acquisition

Image acquisition was controlled via the Raspberry Pi Zero W using a Python script. The image resolution was set to the maximum value of 2048 × 1536 pixels, which was limited to 3 megapixels by the lens. Due to the mounting position of camera systems 2 and 4, the images were rotated 180 degrees to ensure correct image alignment. The images were then triggered. Four images were taken and saved by the camera systems throughout the day. The recordings were carried out at 9:00 a.m., 11:00 a.m., 1:00 p.m., and 3:00 p.m. The trial period of the recording extended from 9 August to 7 December 2023.

2.3. Image/Data Pre-Processing

The recorded and saved images are available as raw data. Before they are used for model calculation, some data pre-processing steps are necessary, such as annotation and augmentation of the images as well as appropriate dataset splitting for model training, testing, and validation.

2.3.1. Image Annotation

Object recognition involves class identification and localisation in an image. In a supervised learning setting, annotations are used to tell the model which of the growing, flowering, and fruiting classes is present and where it is located in the image. During object recognition, a rectangular bounding box is drawn around the object to define its position. A class assignment is then made. These metadata have a clear assignment and serve the model as the basis for the prediction.
The annotation was carried out using Roboflow [26], a freely accessible web application that provides tools for annotation and comprehensive calculations for tasks in the field of computer vision. The images were uploaded to Roboflow and the annotation tool was used to mark the area of the image where the chilli plant was located with a bounding box. The number of bounding boxes was based on the number of objects in the image. Figure 4 illustrates the labelling of an image with the categories “flowering” and “growing”. Roboflow also converts the images to 640 × 480 in order to optimise computing power and retain a sufficient level of detail.

2.3.2. Image Augmentation

Too little data can lead to underfitting or overfitting of deep learning networks. Augmentation makes it possible to increase the number of existing images in order to provide more training data for the model, increase sample diversity, and improve the robustness and generalisation ability of the training model. This involves altering the images, for example, by mirroring or rotating them. Augmentation is only carried out on the training data. This is to prevent similar images from appearing in the training, test, and validation dataset, as only unmodified images allow a true performance evaluation. It is crucial that the generated bounding boxes and their coordinates are also adapted to the changed conditions. The position of the object within the image can change as a result of the augmentation.
In this study, three transformations were applied:
  • Flipping the image on the horizontal axis.
  • Rotating the image by an angle up to 40 degrees, selected randomly from the uniform distribution.
  • Randomly changing the brightness and contrast of the image.
The open-source library Albumentations [27] was used for implementing these options. The parameter p = 0.5 was set for the probability of the augmentation being applied. Figure 5 shows examples after augmentation. The pipeline was run three times per image and generated a three times larger image dataset with random augmentations from the raw data. Our goal here was to obtain about 1500 images per class, as suggested in [28].

2.3.3. Datasets

The datasets served as the basis for model training. Two datasets for images were created: one dateset from a bird’s-eye view (BV) and another from a bird’s-eye and side views (BSV).
Before augmentation, the image data were divided in a ratio of 70-10-20. As most of the data were required for training the model to learn features and patterns, 70% of the data were used for this. After each complete run of the calculation on the training dataset, the model was tested on the validation dataset. This is to prevent overfitting to the training data. For this purpose, 10% of the data that were not included during training were used. In order to obtain unbiased test results of the trained model, 20% of the data were withheld for the test dataset. As the validation data were already used during training, unused data were crucial for the evaluation. The test evaluation only took place on the test set (BV) for all trained models from the bird’s-eye view and bird’s-eye and side views. The reason for this was the performance comparison of the two models for the data from the bird’s-eye view.
The raw dataset from the bird’s-eye and side views consisted of a total of 1489 images, of which 742 images were from the bird’s-eye view and 747 images from the side view. After augmentation, the bird’s-eye view dataset consisted of a total of 1780 images with 2941 labels (bounding boxes). Since only the training dataset was extended by augmentation, the 88-4-8 training ratio changed to 88-4-8 training, validation, and test images. As can be seen on the left in Figure 6, the growing class contained 1644 labels; flowering, 575; and fruiting, 722. The original dataset for the bird’s-eye and side views comprised 3573 images. As the test dataset was identical to that of the bird’s-eye view, the total dataset consisted of 3423 images with a total of 5441 labels. The new ratio of training, validation, and test datasets was 90-5-5. The growing, flowering, and fruiting classes contained 2340, 1002, and 2099 labels, respectively, as shown in the bar chart on the right in Figure 6.
The distribution of the labels for the training, validation, and test data depending on the views and the classes growing, flowering, and fruiting can be seen in Table 2.

2.4. Detection and Classification Methods and Tools

2.4.1. Model Training Procedure

The third part of the computer vision pipeline (Figure 1) consists of the calculations of the model. The YOLOv8 model [28], a deep learning model developed based on Ultralytics, was used for the task of recognising the growth stages, as it is currently regarded as the most powerful model for the recognition of objects of full sizes for the following reasons [29]: (a) lightweight network architecture, (b) effective feature fusion methods, and (c) more accurate detection results than its predecessor versions YOLOv3–YOLOv7. The training process of the growth stage detection procedure is shown Figure 7.
The YOLOv8 network structure consists of three main parts (see Figure 8 and Figure 9):
  • The backbone is basically the same as that of YOLOv5 and consists of a series of convolutional and deconvolutional layers to extract features. It also incorporates residual connections and bottleneck structures to reduce network size and improve performance [30]. This part uses the C2f module as the basic building block that replaces the C3 module in YOLOv5. This offers better gradients to support different model scales by adjusting the number of channels. At the end of the backbone, an SPPF module is used, and three Maxpools of size 5 × 5 are passed serially, and then, each layer is concatenated, so as to guarantee the accuracy of objects at various scales while ensuring a light weight simultaneously [29].
  • The neck part uses multi-scale feature fusion techniques to merge feature maps from different stages of the backbone to improve feature representation capabilities. The feature fusion method used by YOLOv8 is still PAN-FPN [31,32], which strengthens the fusion and utilisation of feature layer information at different scales. Two upsampling and multiple C2f modules together with the final decoupled head structure are used to compose the neck module. The idea of decoupling the head used in YOLOvx is used by YOLOv8 for the last part of the neck. This combines confidence and regression boxes to achieve a new level of accuracy [29].
  • The head, which is responsible for the final object detection and classification tasks, adopts a decoupled head structure, separating the classification and detection head branches. The detection head consists of a series of convolutional and deconvolutional layers to generate detection results, while the classification head uses global average pooling to classify each feature map. It also adopts an anchor-free strategy, abandoning the anchor boxes used in YOLOv7, which reduces the number of box predictions and improves the speed of non-maximum suppression (NMS) [30]. For loss computation, YOLOv8 uses the task-aligned assigner positive sample assignment strategy. It uses BCE loss for the classification branch and distribution focal loss (DFL) and CIoU loss for loss computation in the regression branch. YOLOV8 requires decoding the integral representation of bbox shapes in distribution focal loss, using Softmax and Conv computations to transform them into conventional 4-dimensional bboxes. The head section outputs feature maps at six different scales. The predictions from the classification and bbox branches at three different scales are concatenated and dimensionally transformed. The final output includes three feature maps at scales of 80 × 80 , 40 × 40 , and 20 × 20 (inputs of the “detect” blocks in Figure 8).

2.4.2. YOLO Variants and Model Parameters

YOLOv8 is available in five variants, which differ mainly in the number of convolutional layers: YOLOv8n (nano), YOLOv8s (small), YOLOv8m (medium), YOLOv8l (large), and YOLOv8x (extra large). In this study, the three variants YOLOv8n, v8m, and v8l are used to analyse the two datasets (see Section 2.3.3). Smaller models, which are represented by the YOLO variants, require less computing power for the prediction of unknown data. Larger architectures, such as YOLOv8l, are more accurate but require more computing power. The present results should serve as a guide in the future to better understand the required computing power depending on the accuracy requirements.
There are a large number of parameters for customising and controlling the YOLO models, which can be modified on an individual basis. These include organisational aspects, such as the management of the storage location and the naming of the project, as well as hyperparameters that can have a decisive influence on the accuracy and speed of the trained model.
The calculations in this work are largely carried out using the standard parameters. Table 3 shows the most important hyperparameters for the calculations:
  • The number of epochs indicates how often the entire training dataset was run through the YOLOv8 algorithm. A value of 300 epochs was selected to ensure adaptation on the one hand and not to overstretch the calculation time on the other.
  • The batch size specifies how many parts the training dataset is divided into. After each run, the model performance is adjusted to improve the learning performance. One epoch corresponds to the run of all batches. A value of 16 batches was selected for the application, as smaller batches lead to faster convergence, but are subject to greater fluctuations. More batches lead to more accurate estimates, but require more storage capacity, and therefore, computing power.
  • The image size refers to the dimensions of the images in width and height that are processed by the model. Higher resolutions require more computing resources, smaller resolutions can lead to a loss of information. As this task involves small details such as buds, blossoms, and fruit buds, an image format of 640 × 480 was selected. YOLO models process images in a 1:1 ratio, so the shorter side of the image was filled with black bars to maintain the original ratio.
The models obtained were pre-trained on the COCO dataset [33]. These models have already developed the ability to recognise features and patterns. The COCO dataset, which contains a total of 80 classes, also includes the “potted plant” class. In order to compare the models but stay within the limits of computational resources, the YOLOv8n, v8m, and v8l models were used.

2.4.3. Computation Environment

The setup for the calculation consists of a Dell Precision 3660 workstation with central processing unit (CPU) Intel Core i9-13900, a random access memory (RAM) of 64 GB, and a graphical processing unit (GPU) from NVIDIA RTX A5000, as can be seen in Table 4. The calculations are performed in the integrated development environment (IDE) Visual Studio Code; Python version 3.11.5 is used.

2.5. Performance Evaluation Metrics

There are different approaches for evaluating object recognition models. The intersection over union ( I o U ) is a value between 0 and 1 that describes the overlap between the predicted bounding box and the ground truth, the actual bounding box, as shown in the formula [34]
I o U = a r e a   o f   o v e r l a p a r e a   o f   u n i o n
Sustainability 16 06420 i001
An IoU value of 1 means complete overlap, while a value of 0 indicates that the boxes do not touch at all. Each bounding box that is recognised can be assigned to one of the four components listed in Table 5.
The classification results of each image can be visualised using a confusion matrix. This matrix illustrates the extent to which the predicted classes correspond to the actual ground truth classes. Values such as precision, recall and F1 score can be derived from this.
Precision (P) evaluates the proportion of correct positive predictions in relation to all positive predictions and shows the ability of the model to minimise false positive results, as the following formula shows:
P = T P T P + F P
Recall (R) is the proportion of correct positive predictions in relation to all actual positive instances and evaluates the ability of the model to recognise all instances of a class. This is calculated as follows:
R = T P T P + F N
The precision–recall curve (PR curve) is a graph that shows precision values on the y-axis and recall values on the x-axis. High precision and recall values are a positive indication of the model. Average precision ( A P ) is the area under the PR curve. Mean average precision ( m A P ) is the average of the AP values across all classes and is calculated as follows:
m A P = 1 C i = 1 C A P i
where A P i denotes the AP value for the ith class, and C the total number of classes contained in the dataset. This is useful in scenarios with multi-class object recognition to enable a comprehensive evaluation of model performance [34].
Loss functions play a decisive role in the context of training a model for object recognition. These functions are used to determine the difference between the predicted bounding boxes and the actual annotations. They provide information on how effectively the model learns during training. Typical components of loss functions are the bounding box loss (box_loss), the class loss (cls_loss), and defocus loss (dfl_loss). box_loss takes into account the error in the coordinates of bounding boxes in the prediction. The aim is for the model to match the predicted bounding boxes to the ground truth boxes. cls_loss measures the deviation in the prediction of the object class for each bounding box. This ensures precise identification of the class of the object. dfl_loss is used to optimise object recognition with blurred or defocused images [35].

3. Experimental Results and Comparative Analysis

The last step of the computer vision pipeline in Figure 1 involves evaluating the models. In this section, the results of YOLOv8n, YOLOv8m, and YOLOv8l are presented using the bird’s-eye view and bird’s-eye and side view datasets. The metrics described in Section 2.5 serve as benchmarks for comparison between the models.

3.1. Image Recordings

The raw dataset of all image recordings comprises 1489 images that were generated with the four camera systems. The data collection period spanned four months, with the camera systems generating the images throughout the day (see Section 2.1). Figure 10 presents images of the growth stages—from left to right, the growing, flowering, and fruiting phases can be seen from the bird’s-eye view.
The colouring of the images from the side view reflects the prevailing conditions in the grow box. In contrast, the images from the bird’s-eye view, especially during the flowering and fruiting stages of growth, have a yellowish tinge. Figure 11 shows an image of a plant taken on the same day, with a yellowish cast visible on the left-hand side, while the right-hand side is unaffected. This only occurred when the front of the grow box was closed. As the lens settings were made with the grow box open, the camera system did not adapt to the changed lighting conditions. This lack of white balance led to the yellowish discolouration of the images.
The image recordings of all camera systems show a distortion at the edge of the image. This barrel distortion is clearly recognisable in Figure 12, particularly at the left edge of the image on the base of the frame. The course of the frame is curved inwards, which is emphasised by the vertical blue line. At short focal lengths, this distortion occurs mainly with wide-angle lenses. The distance between the camera lens and the plants is determined by the dimensions of the grow box.
Augmenting the images made it possible to enlarge the dataset. With large rotations of the images, the bounding boxes are expanded, as can be seen on the right of Figure 13. In contrast, the bounding box on the unprocessed left-hand image is closely aligned with the canopy of the plant.

3.2. Model Training Results

All calculations were carried out under the hardware conditions described in Section 2.4.3. The evaluation metrics of the training results are the loss functions and the m A P values, as described in Section 2.5. The loss functions provide information about the inaccuracy of the predictions. The m A P values provide insights into the classification and localisation of the objects in the model. Both the m A P 50 and m A P 50 95 values are considered here, whereby these differ in their I o U threshold value. The m A P 50 value refers to the calculation of an I o U threshold value of 0.5, while the m A P 50 95 value takes into account various I o U threshold values from 0.5 to 0.95 in steps of 0.05. The test results are additionally described with a confusion matrix and precision and recall are given.
For clear identification, the models with the bird’s-eye view training dataset are labelled with BV (YOLOv8n-BV, YOLOv8m-BV, YOLOv8l-BV) and those with the side and bird’s-eye views with BSV (YOLOv8n-BSV, YOLOv8m-BSV, YOLOv8l-BSV).

3.2.1. Training Results: Bird’s-Eye View

The YOLOv8n-BV model consists of 3.01 million parameters and 225 layers. Under the hardware conditions described, training for all 300 epochs takes 35 min. The finished model size is 6.3 MB. Figure 14 illustrates the overall results of the model calculation over all epochs, with the loss functions during training and validation, the m A P 50 and m A P   50 95 , as well as the precision and recall curve. The graphs of the loss functions decrease steadily over the training epochs. A sudden drop in the d f l _ l o s s and b o x _ l o s s can be recognised in the last 10 epochs. The validation curves drop sharply at the beginning, and then, stabilise at a low level. The m A P 50 curve converges after a few epochs and the m A P 50 95 graph shows an exponential convergence.
The m A P value of the model on the validation data is 99.0% and the m A P 50 95 value is 94.4%. Table 6 shows the distribution of the m A P values across the three classes growing, flowering, and fruiting. The m A P 50 values of all classes are very high and are separated by a maximum of 1.0%. Flowering and fruiting achieve higher m A P 50 95 values (96.6% and 96.5%) than growing (90.1%).
Since the training of YOLOv8m-BV did not show any improvements over the last 50 epochs of validation losses, the training stopped early after 197 epochs. The calculation for the included 25.85 million parameters and 295 layers thus took 1 h and 2 min. The model size of YOLOv8m-BV is 52 MB. The graphs of the m A P values in Figure 15 show similar curves to YOLOv8n-BV across all epochs. The loss functions of the training decrease sharply after the first epochs, and then, change to a constant decrease. On the validation data, the losses also strongly decrease at the beginning, and then, converge to a minimum.
The YOLOv8m-BV model achieves an m A P 50 value of 99.0% and an m A P 50 95 value of 94.2% across all classes on the validation dataset, as can be seen in Table 7. With regard to the m A P 50 , the values of the individual classes show only slight differences, with growing achieving the lowest value of 98.0%. Growing also achieves the lowest value for m A P 50 95 at 89.9%, while fruiting achieves the highest value at 97.0%.
The validation of the model across all classes achieves an m A P 50 value of 98.9% and an m A P 50 95 value of 94.8%. Looking at the individual classes, flowering and fruiting have the same m A P 50 value of 99.5% and growing 97.8%. Growing, flowering, and fruiting achieve a m A P 50 95 values of 91.0%, 95.8%, and 97.6%, respectively, as can be seen in Table 8.
With 43.63 million parameters and 365 layers, the YOLOv8l-BV is the largest model architecture in this study. The training was stopped prematurely after 1:57 h and ran over 251 epochs. The best model was selected in epoch 201 and had a size of 87.7 MB. Figure 16 shows that the loss functions decrease continuously during training and converge to a minimum early on the validation dataset. The m A P 50 graph converges to a maximum after a few epochs. The course of the m A P 50 95 approaches the maximum exponentially.

3.2.2. Training Results: Bird’s-Eye and Side Views

As the same YOLO models were used for datasets BV and BSV, the number of parameters, the layers, and the model sizes do not differ. This enables a direct comparison of the results between the two datasets and makes it easier to analyse the performance of the YOLO models.
The loss functions of YOLOv8n-BSV during training, as can be seen in the first three graphs of Figure 17, fall sharply monotonically at the beginning, and then, transition to a moderate drop after about 30 epochs. The loss functions of the validation converge after a rapid initial decrease. The c l s _ l o s s graph shows a minimal slope in the last 30 epochs. The training lasted 54 min and stopped early after 230 epochs, as the validation loss of the last 50 epochs did not change. The best results were observed in epoch 180. The graph of the m A P 50 rises sharply at the beginning and flattens out slightly after reaching the maximum, which becomes apparent when the y-axis is set precisely. The m A P 50 95 resembles a logarithmic curve.
The m A P 50 value is 98.8% and is almost identical in the respective individual classes, with flowering achieving the lowest value at 97.8%. Growing has the lowest m A P 50 95 value at 89.1%. Flowering and fruiting are almost identical at 95.1% and 94.9% and achieve a total of 93.0%, as shown in Table 9.
After 2 h and 28 min, the training of YOLOv8m-BSV was finished early over 241 epochs. The best model was achieved in epoch 191. In Figure 18, the training losses show strongly decreasing curves at the beginning, and then, approach a minimum. The validation losses c l s _ l o s s and d f l _ l o s s increase slightly towards the end of training. The m A P 50 graph rises sharply and flattens out slightly, which becomes apparent with precise scale adjustment. The m A P 50 95 graph also rises sharply at the beginning, but then slows down and converges in the last epochs.
The m A P 50 value of the trained model on the validation data of the entire classes is 98.8%. The m A P 50 95 value is slightly lower at 94.4%, as shown in Table 10. Growing achieves the highest score of 99.4% for the m A P 50 and the lowest score of 91.6% for the m A P 50 95 . Flowering achieved the highest value for the m A P 50 95 with 96.6%.
Figure 19 shows the training values of the YOLOv8l-BSV, which stopped prematurely after 272 of 300 epochs. The training thus lasted 4 h and 8 min. The training loss graphs fall sharply at the beginning, and then, transition to a moderate decline. The validation losses also drop sharply at the beginning, slow down after about 50 epochs, and then, approach their minimum. A slight increase in d f l _ l o s s can be seen after 100 epochs. The m A P 50 and m A P 50 95 graphs behave in a similar way to YOLOv8n-BSV.
Table 11 illustrates the values achieved by YOLOv8l-BSV. The m A P 50 value is 98.0% and the m A P 50 95 value is 93.8%. An analysis of the individual classes shows that the growing and fruiting classes are almost the same for m A P 50 . Only flowering achieves a slightly lower value of 96.9%. Growing achieves the lowest value with an m A P 50 95 of 91.7% and flowering the highest at 95.8%.

3.2.3. Summary of the Training Results

The performance of the proposed plant growth stage detection system using the YOLOv8n-BV, YOLOv8m-BV, and YOLOv8l-BV models was high. The loss functions of the three models indicate improved model learning, as the values of the training losses continuously decrease and those of the validation losses do not increase. This trend indicates that no overfitting has taken place. YOLOv8n-BV shows a sudden drop in training losses at the end of training. This behaviour can be interpreted in the context of the closure of the mosaic augmentation. In the last 10 epochs, the augmentation of the data is closed in order to avoid a deterioration of the training performance. The training losses of YOLOv8n-BSV, YOLOv8m-BSV, and YOLOv8l-SBV decrease over all epochs. The validation losses c l s _ l o s s for YOLOv8n-BSV and YOLOv8m-BSV and d f l _ l o s s for YOLOv8m-BSV and YOLOv8l-BSV increase slightly, indicating the beginning of overfitting of the models.
Table 12 summarises the training results for all the models. The models achieve similarly high m A P 50 values, with YOLOv8n-BV and YOLOv8m-BV achieving the highest accuracies of 99.0% each and YOLOv8l-BSV achieving the lowest value of 98.0%. With dataset BV, YOLOv8l-BV achieves the highest value for m A P 50 95 with 94.8%, while YOLOv8n-BSV achieves the lowest value of 93.0% with dataset BSV. All the models, except YOLOv8n-BV, stop training prematurely. YOLOv8m-BV is the first to stop the calculation, after 197 epochs. With the BSV dataset, all the models stop the training earlier.
The total training time for all the models is 11 h and 4 min. As expected, the small model YOLOv8n-BV with the smaller dataset BV has the shortest training time of 38 min. The YOLOv8l-BSV takes the longest to train at 4 h and 8 min.

3.3. Test Results

Test dataset BV, consisting of 149 images and 242 labels, was used for all the trained models. This ensures the comparability of all the models. For the interpretation of the results, the m A P 50 95 of all classes is shown on the one hand and the confusion matrix is used on the other hand, as it provides a clear representation of the type of error. Precision and recall can be derived from this.

3.3.1. Test Results Bird’s-Eye View

The m A P 50 95 values of YOLOv8n-BV, YOLOv8m-BV, and YOLOv8l-BV are listed in Table 13. It contains both the results of the individual classes and the overall values. YOLOv8l-BV achieves the highest m A P 50 95 value across all classes on test dataset BV with 93.4%. YOLOv8m-BV achieves the lowest value with 0.8% less. YOLOv8n-BV achieves 92.9% and lies between the medium and large models. All three models achieve almost similar results for the growing and fruiting classes. YOLOv8l-BV achieves the highest values. The largest model also achieves the highest accuracy for the flowering class with 96.1%. In summary, YOLOv8l-BV has the highest m A P values overall in all classes, indicating precise object recognition and localisation, especially in the flowering stage.
In the context of object recognition, the confusion matrix provides insight into the accuracy of predictions and the frequency of misclassifications. The “background” class is an essential component of the YOLO algorithm and is automatically output, classifying regions without objects. Figure 20 shows the confusion matrices of the YOLOv8n-BV, YOLOv8m-BV, and YOLOv8l-BV test results. The true classes are shown in the columns, the predicted classes of the calculation are shown in the rows. Table 2 (in Section 2.3.3) shows that the test dataset BV contains 242 labels.
YOLOv8m-BV predicts 239 of the labels which also correspond to the ground truth, as shown in the confusion (matrix (a)). The model thus achieves the highest number of TP values. YOLOv8m-BV (see matrix (b)) and YOLOv8l-BV (see matrix (c)) follow, with 238 TP each. YOLOv8m-BV (b) and YOLOv8l-BV (c) are error free in the fruiting class, recognising all labels. All the models predict fruiting once, although it corresponds to the class flowering (row 2, column 1). The same is true for two FNs in the growing class, which are misinterpreted twice as background (row 4, column 3). Each model predicts two labels in the growing class, although it corresponds to the background class (row 3, column 4). YOLOv8m-BV achieves the most frequent FPs with four errors, YOLOv8l-BV the fewest with two errors.

3.3.2. Test Results Bird’s-Eye and Side View

Table 14 shows the training results of YOLOv8n-BSV, YOLOv8m-BSV, and YOLOv8l-BSV on the test dataset BV. All the models achieved high values on the test dataset. YOLOv8m-BSV achieved the highest m A P 50 95 value of 96.3% and YOLOv8l-BSV achieved the second highest value of 95.6%, 0.7% lower. YOLOv8m-BSV also achieved the highest accuracies in the individual growing, flowering, and fruiting classes. YOLOv8n-BSV achieved the lowest overall result with 93.7%. Looking at the individual classes, it can be seen that all the models achieved particularly good results for the flowering class, with values between 97.3% and 98.8%. The lowest scores were achieved by all the models in the growing class.
As can be seen in Figure 21 of confusion matrix (a), YOLOv8n-BSV achieves the most TPs, with 240 out of the 242 labels from the test dataset BV. This is followed by YOLOv8m-BSV and YOLOv8l-BSV, with 239 true predictions each. YOLOv8m-BSV and YOLOv8l-BSV have exactly the same values. As the models in Section 3.3.1 have already shown, all the models predict fruiting once, even though it corresponds to the flowering class (row 2, column 1). YOLOv8n-BSV has the fewest FNs and FPs with two and one errors, respectively.

3.3.3. Summary of the Test Results

Table 15 shows the m A P 50 95 values achieved by all the trained models on the test dataset BV. These are compared with the m A P 50 95 values after training on the respective validation datasets. To illustrate the validation results by way of example, Figure 22 and Figure 23 are shown.
Overall, all the models achieve a high m A P 50 95 value on the test dataset BV. YOLOv8m-BSV achieves the highest value on the test dataset with 96.3%. With 3.7% less, YOLOv8m-BV is the model with the lowest m A P 50 95 value at 92.6%. Both models achieve m A P 50 95 values in the middle of the range on their respective validation datasets.
The precision and recall values are derived from TPs, FNs, and FPs. The values of all the models are very close to each other. YOLOv8n-BSV achieves the highest precision and recall values of all the models with 99.6% and 99.2%. The lowest precision value is achieved by YOLOv8m-BV with 98.4% and the lowest recall value is achieved by YOLOv8l-BV, also with 98.4%.

4. Discussion

4.1. Image Quality

The successful acquisition of data by the camera system over the entire survey period suggests that the chosen image resolution was sufficient to allow reliable identification of features such as buds, flowers, fruit set, and fruit. This is confirmed by the results, as all trained models showed an accuracy of over 98% in identifying the classes.
The influence of image quality on model training, in terms of barrel distortion, yellowing, and widened bounding boxes due to image rotation, cannot be determined from the results presented. Further research is required to gain a full understanding of the impact of these image quality factors.

4.2. Data and Model Discussion

The results show that all the models are able to identify the three growth stages of the plants. They achieved an accuracy of at least 92.6% on the test dataset BV with respect to the metric m A P 50 95 .
The best training results were achieved by the YOLO models trained only on the bird’s-eye view dataset. This does not confirm the assumption that extending the dataset to include the side view improves training performance. YOLOv8l-BV achieved the highest training values with 94.8% m A P 50 95 , which supports the assumption that larger models achieve better accuracies. The smallest YOLOv8n-BV was just behind with 94.4%. In addition, the model did not converge after 300 epochs. In contrast to YOLOv8l-BV, YOLOv8n-BV shows the potential for improvement by increasing the number of epochs. With regard to the training evaluation, the assumption of higher accuracy with larger models cannot be confirmed and requires further investigation.
The test results paint a mixed picture of accuracy depending on model size and data influence. On average, YOLOv8n-BV, YOLOv8m-BV, and YOLO-V8l-BV have lower m A P 50 95 accuracies than YOLOv8n-BSV, YOLOv8m-BSV, and YOLOv8l-BSV. This indicates that models trained on the extended dataset BSV show better generalisation. Of all the models, the medium model (YOLOv8m-BSV) trained on the BSV dataset performed best, followed by the two large models YOLOv8l-BSV and YOLOv8l-BV. This suggests that larger models tend to perform better, although this assumption cannot be conclusively confirmed as the medium model (YOLOv8m-BSV) performed best in this particular application.
Looking at the individual classes in more detail, it is noticeable that all the models achieved the lowest m A P 50 95 values in the growing class, even though this class had the most labels (see Table 2). This observation may be due to the quality of the dataset. It is possible that the recording of four images per day and additional augmentation resulted in some redundancy in the data. This may lead to overfitting of the class during training and affect the ability of the models to generalise to previously unseen data.
The precision and recall values of all the models on the test dataset BV are very close to each other at a high level. Precision plays a crucial role in this study, as accurate prediction of classes or growth stages should ensure that future environmental conditions for plant growth are optimised. The smallest model with the BSV dataset (YOLOv8n-BSV) gave the best results. However, due to the small differences between the values, no general statement can be made about the effect of datasets and model sizes on accuracy.
In particular, the evaluation of the test results showed incorrect predictions corresponding to the background class. These errors could be due to the lack of background images in the dataset. In order to minimise these FP errors and improve the generalisability of the model, it would be advisable to include a dataset with background images in the next calculations. In particular, these should include images of the growth box and future application areas.
It is important to emphasise that the models developed and the results obtained in this work are only valid for the specific use case with the considered plants in the defined grow box environment. The future challenge is to ensure that the developed models can prove their performance in different scenarios and that their robustness and applicability are validated in a broader context.

5. Conclusions and Future Directions

In this work, images of chilli plants of the species C. annuum, grown in a hydroponic environment, were taken from two views over their life cycle, from seedling to fruit. The bird’s-eye and side view data were collected over a period of four months and used to train the models. The three models YOLOv8n, YOLOv8m, and YOLOv8l were trained using the datasets BV and BSV. This resulted in the six models YOLO-V8n-BV, YOLO-V8m-BV, YOLO-V8l-BV, YOLO-V8n-BSV, YOLO-V8m-BSV, and YOLO-V8l-BSV.
All six models are able to recognise the growth stages of chilli plants from a bird’s-eye view. The high accuracy of the models confirms the objective. The HQ Raspberry Pi camera with a 6 mm wide angle lens provided images of sufficient quality to train the YOLO models.
The test results of all the models showed a comparably high level. Overall, the BSV dataset showed the best results in terms of m A P 50 95 and precision. The influence of the model size is not clear, as the medium architecture of YOLO gave the best results.
In order to improve the models in the future, the influence of image quality could be analysed. The short distance between the lens and the plants led to distortion of the images. In addition, part of the dataset has a yellowish tinge. These images could be corrected using image processing algorithms. It is necessary to check whether the corrected images have any effect on the calculation results. In general, these phenomena can be avoided by using appropriate software and high-quality hardware. In addition, the dataset should be checked for redundant or inconspicuous features and compared with the results obtained.
Further optimisation could be achieved by hyperparameter tuning. By increasing the number of epochs, the YOLOv8- BV could achieve better results. In addition, the image resolution (image size) could be increased and a k-fold cross-validation could be performed. Expanding the dataset to include background images could also lead to improvements and could be compared with the results obtained in this work.
The next step is to test the generalisability of the model on unknown data. This could be achieved by compiling a common set of data and evaluating the accuracy of the predictions. The reliability of the system could also be tested by growing new chilli plants. The newly generated images could directly contribute to the improvement of the system by increasing the diversity of the dataset.
In this work, the plant in the images was considered as a whole and fed into the model training. Further research could create a dataset with specific annotations of the buds, flowers, and fruits and compare the training results.
The trained models could be used as a holistic image processing system for targeted and intelligent control of the next cultivation process. Depending on the predictions, growth parameters such as light and nutrient composition could be adjusted. The main objectives of this study were energy consumption and yield, which could be compared with cultivation without an image processing system.
Another option that could be analysed is the need for computing power. Can edge computing, such as Raspberry Pi or Raspberry Pi Zero W, be used, or is centralised computing, such as cloud computing, be required? As larger models require more computing resources, the performance needs to be evaluated as a function of the processing speed or inference time of the models. These devices could be used in a mobile design in greenhouses or vertical indoor farms.
Other applications of machine vision in hydroponics include disease and nutrient deficiency detection, as well as flower and fruit detection for pollination and targeted harvesting. The YOLO models created can be used as pre-trained models and as a starting point for the development of new computations in these application areas.

Author Contributions

Conceptualization, M.J. and J.S.; methodology, F.S. and J.S.; software, F.S.; validation, M.J. and J.S.; formal analysis, F.S.; investigation, F.S. and J.S.; resources, F.S. and J.S.; data curation, F.S. and J.S.; writing—original draft preparation, F.S., M.J. and J.S.; writing—review and editing, F.S., M.J. and J.S.; visualization, F.S., M.J. and J.S.; supervision, M.J.; project administration, M.J. and J.S.; funding acquisition, M.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the German Federal Ministry for Economic Affairs and Climate Action (Bundesministerium für Wirtschaft und Klima—BMWK) grant number ZIM/KK5149908SA3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors are very grateful to Loui Al-Shrouf, who contributed to the acquisition of the funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APAverage precision
BVBird’s-eye view
BSVBird’s-eye and side views
CAISACologne Lab for Artificial Intelligence and Smart Automation
cls lossClass loss
CNNConvolutional neural network
CPUCentral processing unit
dfl lossDefocus loss
DLDeep learning
ECElectric conductivity
FNFalse negative
FPFalse positive
GBGigabyte
GPUGraphical processing unit
IDEIntegrated development environment
IoUIntersection over union
LEDLight-emitting diode
mAPMean average precision
MBMegabyte
MLMachine learning
MVMachine vision
NFTNutrient film technique
PARPhotosynthetically active radiation
pHPotential hydrogen
PPFDPhotosynthetically active photon flux density
PRPrecision–recall
SVSide view
TNTrue negative
TPTrue positive
VIFVertical indoor farming
YOLOYou Only Look Once

References

  1. Despommier, D.D. The Vertical Farm: Feeding the World in the 21st Century; Picador: New York, NY, USA, 2020. [Google Scholar]
  2. Polsfuss, L. PFLANZEN. Available online: https://pflanzenfabrik.de/hydroponik/pflanzen/ (accessed on 5 April 2024).
  3. Chilisorten. Available online: https://chili-plants.com/chilisorten/ (accessed on 2 December 2023).
  4. Drache, P. Chili Geschichte, Herkunft und Verbreitung. Available online: https://chilipflanzen.com/wissenswertes/chili-geschichte/ (accessed on 2 December 2023).
  5. Azlan, A.; Sultana, S.; Huei, C.S.; Razman, M.R. Antioxidant, Anti-Obesity, Nutritional and Other Beneficial Effects of Different Chili Pepper: A Review. Molecules 2022, 27, 898. [Google Scholar] [CrossRef] [PubMed]
  6. Thiele, R. Untersuchungen zur Biosynthese von Capsaicinoiden—Vorkommen und Einfluss von Acyl-Thioestern auf das Fettsäuremuster der Vanillylamide in Capsicum spp. Ph.D. Thesis, Bergische Universität Wuppertal, Wuppertal, Germany, 2008. Available online: https://elekpub.bib.uni-wuppertal.de/urn/urn:nbn:de:hbz:468-20080466 (accessed on 2 December 2023).
  7. Meier, U. Entwicklungsstadien Mono-und Dikotyler Pflanzen: BBCH Monografie, Quedlinburg. 2018. Available online: https://www.openagrar.de/receive/openagrar_mods_00042352 (accessed on 2 December 2023).
  8. Feldmann, F.; Rutikanga, A. Phenological growth stages and BBCHidentification keys of Chilli (Capsicum annuum L., Capsicum chinense JACQ., Capsicum baccatum L. J. Plant Dis. Prot. 2021, 128, 549–555. [Google Scholar] [CrossRef]
  9. Paul, N.C.; Deng, J.X.; Sang, H.-K.; Choi, Y.-P.; Yu, S.-H. Distribution and Antifungal Activity of Endophytic Fungi in Different Growth Stages of Chili Pepper (Capsicum annuum L.) in Korea. Plant Pathol. J. 2012, 28, 10–19. [Google Scholar] [CrossRef]
  10. Paul, A.; Nagar, H.; Machavaram, R. Utilizing Fine-Tuned YOLOv8 Deep Learning Model for Greenhouse Capsicum Detection and Growth Stage Determination. In Proceedings of the 2023 7th International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud), Kirtipur, Nepal, 11–13 October 2023; pp. 649–656. Available online: https://ieeexplore.ieee.org/document/10290335 (accessed on 2 December 2023).
  11. Kamilaris, A.; Prenafeta-Boldú, F.X. A review of the use of convolutional neural networks in agriculture. J. Agric. Sci. 2018, 156, 312–322. [Google Scholar] [CrossRef]
  12. Tian, H.; Wang, T.; Liu, Y.; Qiao, X.; Li, Y. Computer vision technology in agricultural automation—A review. Inf. Process. Agric. 2020, 7, 1–19. [Google Scholar] [CrossRef]
  13. Lin, K.; Chen, J.; Si, H.; Wu, J. A Review on Computer Vision Technologies Applied in Greenhouse Plant Stress Detection. In Advances in Image and Graphics Technologies; IGTA 2013; Communications in Computer and Information Science; Tan, T., Ruan, Q., Chen, X., Ma, H., Wang, L., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; Volume 363. [Google Scholar] [CrossRef]
  14. Wijanarko, A.; Nugroho, A.P.; Kusumastuti, A.I.; Dzaky, M.A.F.; Masithoh, R.E.; Sutiarso, L.; Okayasu, T. Mobile mecavision: Automatic plant monitoring system as a precision agriculture solution in plant factories. IOP Conf. Ser. Earth Environ. Sci. 2021, 733, 012026. [Google Scholar] [CrossRef]
  15. Samiei, S.; Rasti, P.; Ly Vu, J.; Buitink, J.; Rousseau, D. Deep learning-based detection of seedling development. Plant Methods 2020, 16, 103. [Google Scholar] [CrossRef] [PubMed]
  16. Yeh, Y.-H.F.; Lai, T.-C.; Liu, T.-Y.; Liu, C.-C.; Chung, W.-C.; Lin, T.-T. An automated growth measurement system for leafy vegetables. Biosyst. Eng. 2014, 117, 43–50. [Google Scholar] [CrossRef]
  17. Nugroho, A.P.; Fadilah, M.A.N.; Wiratmoko, A.; Azis, Y.A.; Efendi, A.W.; Sutiarso, L.; Okayasu, T. Implementation of crop growth monitoring system based on depth perception using stereo camera in plant factory. IOP Conf. Ser. Earth Environ. Sci. 2020, 542, 012068. [Google Scholar] [CrossRef]
  18. Phänotypisierung. Available online: https://www.pflanzenforschung.de/de/pflanzenwissen/lexikon-a-z/phaenotypisierung-10020 (accessed on 2 December 2023).
  19. Li, Z.; Guo, R.; Li, M.; Chen, Y.; Li, G. A review of computer vision technologies for plant phenotyping. Comput. Electron. Agric. 2020, 176, 105672. [Google Scholar] [CrossRef]
  20. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  21. Hespeler, S.C.; Nemati, H.; Dehghan-Niri, E. Non-destructive thermal imaging for object detection via advanced deep learning for robotic inspection and harvesting of chili peppers. Artif. Intell. Agric. 2021, 5, 102–117. [Google Scholar] [CrossRef]
  22. Coleman, G.R.; Kutugata, M.; Walsh, M.J.; Bagavathiannan, M. Multi-growth stage plant recognition: A case study of Palmer amaranth (Amaranthus palmeri) in cotton (Gossypium hirsutum). arXiv 2023, arXiv:2307.15816. [Google Scholar] [CrossRef]
  23. Zhang, P.; Li, D. CBAM + ASFF-YOLOXs: An improved YOLOXs for guiding agronomic operation based on the identification of key growth stages of lettuce. Comput. Electron. Agric. 2022, 203, 107491. [Google Scholar] [CrossRef]
  24. Xiao, F.; Wang, H.; Xu, Y.; Shi, Z. A Lightweight Detection Method for Blueberry Fruit Maturity Based on an Improved YOLOv5 Algorithm. Agriculture 2023, 14, 36. [Google Scholar] [CrossRef]
  25. grow-shop24. DiamondBox Silver Line SL150, 150×150×200 cm. Available online: https://www.grow-shop24.de/diamondbox-silver-line-sl150 (accessed on 8 June 2023).
  26. Build Vision Models with Roboflow. Available online: https://docs.roboflow.com/ (accessed on 6 December 2023).
  27. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and Flexible Image Augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef]
  28. Jocher, G.; Chaurasia, A.; Qiu, J. YOLO by Ultralytics. Available online: https://github.com/ultralytics/ultralytics (accessed on 30 November 2023).
  29. Lou, H.; Duan, X.; Guo, J.; Liu, H.; Gu, J.; Bi, L.; Chen, H. DC-YOLOv8: Small-Size Object Detection Algorithm Based on Camera Sensor. Electronics 2023, 12, 2323. [Google Scholar] [CrossRef]
  30. Pan, Y.; Xiao, X.; Hu, K.; Kang, H.; Jin, Y.; Chen, Y.; Zou, X. ODN-Pro: An Improved Model Based on YOLOv8 for Enhanced Instance Detection in Orchard Point Clouds. Agronomy 2024, 14, 697. [Google Scholar] [CrossRef]
  31. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. Available online: https://ieeexplore.ieee.org/document/8579011 (accessed on 6 April 2024).
  32. Lin, T.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. Available online: https://ieeexplore.ieee.org/document/8099589 (accessed on 6 April 2024).
  33. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8693, pp. 740–755. Available online: https://link.springer.com/chapter/10.1007/978-3-319-10602-1_48 (accessed on 6 April 2024).
  34. Padilla, R.; Passos, W.L.; Dias, T.L.B.; Netto, S.L.; da Silva, E.A.B. A Comparative Analysis of Object Detection Metrics with a Companion Open-Source Toolkit. Electronics 2021, 10, 279. [Google Scholar] [CrossRef]
  35. Akbarnezhad, E. YOLOv8 Projects #1 “Metrics, Loss Functions, Data Formats, and Beyond”. Available online: https://www.linkedin.com/pulse/yolov8-projects-1-metrics-loss-functions-data-formats-akbarnezhad/ (accessed on 16 November 2023).
Figure 1. Structure of the machine vision system from image acquisition and image pre-processing to the modelling process (computer vision).
Figure 1. Structure of the machine vision system from image acquisition and image pre-processing to the modelling process (computer vision).
Sustainability 16 06420 g001
Figure 2. Experimental setup for image generation, with the camera systems outlined in blue.
Figure 2. Experimental setup for image generation, with the camera systems outlined in blue.
Sustainability 16 06420 g002
Figure 3. (a) Experimental set-up with dimensions labelled in black/bold for the growth stage (growing) and in blue/bold for the stages flowering and fruiting; (b) mounted camera systems for the side view (left) and the bird’s-eye view (right).
Figure 3. (a) Experimental set-up with dimensions labelled in black/bold for the growth stage (growing) and in blue/bold for the stages flowering and fruiting; (b) mounted camera systems for the side view (left) and the bird’s-eye view (right).
Sustainability 16 06420 g003
Figure 4. Annotation of an image with the classes “flowering” (red boxes) and “growing” (purple box) using Roboflow 100.
Figure 4. Annotation of an image with the classes “flowering” (red boxes) and “growing” (purple box) using Roboflow 100.
Sustainability 16 06420 g004
Figure 5. The same picture after applying the transformations three times with the bounding box.
Figure 5. The same picture after applying the transformations three times with the bounding box.
Sustainability 16 06420 g005
Figure 6. Data distribution of the labels according to the classes growing, flowering, and fruiting for (a) the bird’s-eye view; (b) the bird’s-eye and side views.
Figure 6. Data distribution of the labels according to the classes growing, flowering, and fruiting for (a) the bird’s-eye view; (b) the bird’s-eye and side views.
Sustainability 16 06420 g006
Figure 7. Training flow of the developed growth stage detection procedure.
Figure 7. Training flow of the developed growth stage detection procedure.
Sustainability 16 06420 g007
Figure 8. YOLOv8 network structure. Legend: CBS—convolution + batch normalisation + sigmoid-weighted linear unit (SiLU); C2f—CSP (cross-stage partial) bottleneck with 2 convolutions, fast version; SPPF—spatial pyramid pooling, fast version; Concat—concatenation; Detect: detector; FPN—feature pyramid network; PAN—path aggregation network. See Figure 9 for representation of the sub-blocks used.
Figure 8. YOLOv8 network structure. Legend: CBS—convolution + batch normalisation + sigmoid-weighted linear unit (SiLU); C2f—CSP (cross-stage partial) bottleneck with 2 convolutions, fast version; SPPF—spatial pyramid pooling, fast version; Concat—concatenation; Detect: detector; FPN—feature pyramid network; PAN—path aggregation network. See Figure 9 for representation of the sub-blocks used.
Sustainability 16 06420 g008
Figure 9. YOLOv8 architecture sub-blocks used in Figure 8. Legend: SiLU—sigmoid-weighted linear unit; Conv2d—2D convolution; BatchNorm2d—2D batch normalisation; Maxpool2d—2D max pooling; box_loss—(bounding) box loss; cls_loss—class loss.
Figure 9. YOLOv8 architecture sub-blocks used in Figure 8. Legend: SiLU—sigmoid-weighted linear unit; Conv2d—2D convolution; BatchNorm2d—2D batch normalisation; Maxpool2d—2D max pooling; box_loss—(bounding) box loss; cls_loss—class loss.
Sustainability 16 06420 g009
Figure 10. Images of the 3 growth stages from the bird’s-eye view.
Figure 10. Images of the 3 growth stages from the bird’s-eye view.
Sustainability 16 06420 g010
Figure 11. Left image with yellow tint and right image without yellow tint.
Figure 11. Left image with yellow tint and right image without yellow tint.
Sustainability 16 06420 g011
Figure 12. Image distortion with straight blue bar.
Figure 12. Image distortion with straight blue bar.
Sustainability 16 06420 g012
Figure 13. Bounding box after rotation of the image.
Figure 13. Bounding box after rotation of the image.
Sustainability 16 06420 g013
Figure 14. Loss functions and m A P values after 300 epochs of the YOLOv8n-BV model with the dataset BV.
Figure 14. Loss functions and m A P values after 300 epochs of the YOLOv8n-BV model with the dataset BV.
Sustainability 16 06420 g014
Figure 15. Loss functions and m A P values after 300 epochs of the YOLOv8m-BV model with the dataset BV.
Figure 15. Loss functions and m A P values after 300 epochs of the YOLOv8m-BV model with the dataset BV.
Sustainability 16 06420 g015
Figure 16. Loss functions and m A P values after 300 epochs of the YOLOv8l-BV model with the dataset BV.
Figure 16. Loss functions and m A P values after 300 epochs of the YOLOv8l-BV model with the dataset BV.
Sustainability 16 06420 g016
Figure 17. Loss functions and m A P values after 300 epochs of the YOLOv8n-BSV model with the dataset BSV.
Figure 17. Loss functions and m A P values after 300 epochs of the YOLOv8n-BSV model with the dataset BSV.
Sustainability 16 06420 g017
Figure 18. Loss functions and m A P values after 300 epochs of the YOLOv8m-BSV model with the dataset BSV.
Figure 18. Loss functions and m A P values after 300 epochs of the YOLOv8m-BSV model with the dataset BSV.
Sustainability 16 06420 g018
Figure 19. Loss functions and m A P values after 300 epochs of the YOLOv8l-BSV model with the dataset BSV.
Figure 19. Loss functions and m A P values after 300 epochs of the YOLOv8l-BSV model with the dataset BSV.
Sustainability 16 06420 g019
Figure 20. Confusion matrices of the test results for the daset BV: (a) YOLOv8n; (b) YOLOv8m; (c) YOLOv8l.
Figure 20. Confusion matrices of the test results for the daset BV: (a) YOLOv8n; (b) YOLOv8m; (c) YOLOv8l.
Sustainability 16 06420 g020
Figure 21. Confusion matrices of the test results for the dataset BSV: (a) YOLOv8n; (b) YOLOv8m; (c) YOLOv8l.
Figure 21. Confusion matrices of the test results for the dataset BSV: (a) YOLOv8n; (b) YOLOv8m; (c) YOLOv8l.
Sustainability 16 06420 g021
Figure 22. Example validation results using YOLOv8m for the growth stage detection from a bird’s-eye view.
Figure 22. Example validation results using YOLOv8m for the growth stage detection from a bird’s-eye view.
Sustainability 16 06420 g022
Figure 23. Example validation results using YOLOv8m for the growth stage detection from bird’s-eye and side views.
Figure 23. Example validation results using YOLOv8m for the growth stage detection from bird’s-eye and side views.
Sustainability 16 06420 g023
Table 1. Components of the camera system used in this study.
Table 1. Components of the camera system used in this study.
ItemQuantityItem Description
14HQ Raspberry Pi camera
246MM WW Raspberry Pi lens
34Raspberry Pi Zero W
41Raspberry Pi 4 B
52Mantona 17996 travel tripod
64USB-A to USB-B-Mini 3 m cable
71USB-C power adapter
85Memory card SanDisk Ultra microSDHC 32 GB
91Energenie Uni-4-fold USB charger
Table 2. Listing of labels by class and view.
Table 2. Listing of labels by class and view.
Bird’s-Eye ViewBird’s-Eye and Side Views
Train (BV) Val (BV) Test B (V) Train (BSV) Val (BSV) Test (BSV)
Growing446771212112107121
Flowering49828499143949
Fruiting624267219329572
Table 3. Information on the hyperparameters for the calculations.
Table 3. Information on the hyperparameters for the calculations.
EpochsBatch SizeImage SizeModel
30016 640 × 480 YOLOv8n
YOLOv8m
YOLOv8l
Table 4. Calculation setup.
Table 4. Calculation setup.
DataValue
CPU12th Gen Intel Core i9-13900
RAM64 GB DDR5
GPUNVIDIA RTX A5000
AlgorithmYOLOv8n, -v8m, -v8l
Table 5. Assignment of recognised bounding boxes with explanation.
Table 5. Assignment of recognised bounding boxes with explanation.
AssignmentExplanation
True Positive (TP)Bounding box in the correct position (positive) and correct prediction (true)
False Positive (FP)Bounding box in the right place (positive), but wrong prediction (false)
False Negative (FN)Bounding box not recognised (negative) and incorrect prediction (false)
True Negative (TN)Bounding box not recognised (negative) and correct prediction (true); no influence for multi-class tasks
Table 6. m A P values of the YOLOv8n-BV model on the validation set V.
Table 6. m A P values of the YOLOv8n-BV model on the validation set V.
ClassmAP 50/%mAP 50–95/%
All99.094.4
Growing98.390.1
Flowering99.396.6
Fruiting99.596.5
Table 7. m A P values of the YOLOv8m-BV model on the validation set V.
Table 7. m A P values of the YOLOv8m-BV model on the validation set V.
ClassmAP 50/%mAP 50–95/%
All99.094.2
Growing98.089.9
Flowering99.495.9
Fruiting99.597.0
Table 8. m A P values of the YOLOv8l-BV model on the validation set V.
Table 8. m A P values of the YOLOv8l-BV model on the validation set V.
ClassmAP 50/%mAP 50–95/%
All98.994.8
Growing97.891.0
Flowering99.595.8
Fruiting99.597.6
Table 9. m A P values of the YOLOv8n-BSV model on the validation set BSV.
Table 9. m A P values of the YOLOv8n-BSV model on the validation set BSV.
ClassmAP 50/%mAP 50–95/%
All98.893.0
Growing99.489.1
Flowering97.895.1
Fruiting99.394.9
Table 10. m A P values of the YOLOv8m-BSV model on the validation set BSV.
Table 10. m A P values of the YOLOv8m-BSV model on the validation set BSV.
ClassmAP 50/%mAP 50–95/%
All98.894.4
Growing99.491.6
Flowering97.896.6
Fruiting99.295.2
Table 11. m A P values of the YOLOv8l-BSV model on the validation set BSV.
Table 11. m A P values of the YOLOv8l-BSV model on the validation set BSV.
ClassmAP 50/%mAP 50–95/%
All98.093.8
Growing98.591.7
Flowering96.995.8
Fruiting98.794.0
Table 12. Key performance indicators of the trained YOLOv8 models. (The best values are highlighted in bold).
Table 12. Key performance indicators of the trained YOLOv8 models. (The best values are highlighted in bold).
ModelmAP 50/%mAP 50–95/%EpochsTraining Time/hModel Size/MB
Bird’s-eye view
YOLOv8n-BV99.094.43000.5766.3
YOLOv8m-BV99.094.21971.03252
YOLOv8l-BV98.994.82511.95687.7
Bird’s-eye and side views
YOLOv8n-BSV98.893.02300.8956.3
YOLOv8m-SBV98.894.42412.47452
YOLOv8l-BSV98.093.82724.13887.7
Table 13. m A P 50 95 values of YOLOv8n-BV, YOLOv8m-BV, and YOLOv8l-BV on the training dataset BV. (The best values are highlighted in bold).
Table 13. m A P 50 95 values of YOLOv8n-BV, YOLOv8m-BV, and YOLOv8l-BV on the training dataset BV. (The best values are highlighted in bold).
YOLOv8n-BVYOLOv8m-BVYOLOv8l-BV
mAP 50–95/%
All92.992.693.4
Growing91.291.292.1
Flowering95.695.496.1
Fruiting91.991.191.9
Table 14. m A P 50 95 values of YOLOv8n-BSV, YOLOv8m-BSV, and YOLO-v8l-BSV on the training dataset BSV. (The best values are highlighted in bold).
Table 14. m A P 50 95 values of YOLOv8n-BSV, YOLOv8m-BSV, and YOLO-v8l-BSV on the training dataset BSV. (The best values are highlighted in bold).
YOLOv8n-BSVYOLOv8m-BSVYOLOv8l-BSV
mAP 50–95/%
All93.796.395.6
Growing91.894.594.4
Flowering97.398.898.3
Fruiting92.095.794.2
Table 15. Key performance figures of the tested YOLOv8 models with comparative values from training. (The best values are highlighted in bold).
Table 15. Key performance figures of the tested YOLOv8 models with comparative values from training. (The best values are highlighted in bold).
ModelmAP 50/% TestmAP 50–95/% ValidationPrecision/%Recall/%
Bird’s-eye view
YOLOv8n-BV92.994.498.898.8
YOLOv8m-BV92.694.298.498.8
YOLOv8l-BV93.494.899.298.4
Bird’s-eye and side views
YOLOv8n-BSV93.793.099.699.2
YOLOv8m-BSV96.394.499.298.8
YOLOv8l-BSV95.693.899.298.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Schneider, F.; Swiatek, J.; Jelali, M. Detection of Growth Stages of Chilli Plants in a Hydroponic Grower Using Machine Vision and YOLOv8 Deep Learning Algorithms. Sustainability 2024, 16, 6420. https://doi.org/10.3390/su16156420

AMA Style

Schneider F, Swiatek J, Jelali M. Detection of Growth Stages of Chilli Plants in a Hydroponic Grower Using Machine Vision and YOLOv8 Deep Learning Algorithms. Sustainability. 2024; 16(15):6420. https://doi.org/10.3390/su16156420

Chicago/Turabian Style

Schneider, Florian, Jonas Swiatek, and Mohieddine Jelali. 2024. "Detection of Growth Stages of Chilli Plants in a Hydroponic Grower Using Machine Vision and YOLOv8 Deep Learning Algorithms" Sustainability 16, no. 15: 6420. https://doi.org/10.3390/su16156420

APA Style

Schneider, F., Swiatek, J., & Jelali, M. (2024). Detection of Growth Stages of Chilli Plants in a Hydroponic Grower Using Machine Vision and YOLOv8 Deep Learning Algorithms. Sustainability, 16(15), 6420. https://doi.org/10.3390/su16156420

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop