Next Article in Journal
Correction: Saqib et al. Antidiarrheal and Cardio-Depressant Effects of Himalaiella heteromalla (D.Don) Raab-Straube: In Vitro, In Vivo, and In Silico Studies. Plants 2022, 11, 78
Previous Article in Journal
Application and Evaluation of the Antifungal Activities of Glandular Trichome Secretions from Air/Sun-Cured Tobacco Germplasms against Botrytis cinerea
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sunpheno: A Deep Neural Network for Phenological Classification of Sunflower Images

1
Laboratory of Genetics, Wageningen University & Research, 6708 PB Wageningen, The Netherlands
2
Department of Information Engineering and Computer Science, University of Trento, 38123 Trento, Italy
3
IABIMO UEDD INTA CONICET, Buenos Aires 1686, Argentina
4
Faculty of Electrical Engineering and Computer Science, Technische Universität Berlin, 10587 Berlin, Germany
*
Author to whom correspondence should be addressed.
Plants 2024, 13(14), 1998; https://doi.org/10.3390/plants13141998
Submission received: 28 June 2024 / Revised: 16 July 2024 / Accepted: 17 July 2024 / Published: 22 July 2024
(This article belongs to the Special Issue Deciphering Plant Molecular Data Using Computational Methods)

Abstract

:
Leaf senescence is a complex trait which becomes crucial for grain filling because photoassimilates are translocated to the seeds. Therefore, a correct sync between leaf senescence and phenological stages is necessary to obtain increasing yields. In this study, we evaluated the performance of five deep machine-learning methods for the evaluation of the phenological stages of sunflowers using images taken with cell phones in the field. From the analysis, we found that the method based on the pre-trained network resnet50 outperformed the other methods, both in terms of accuracy and velocity. Finally, the model generated, Sunpheno, was used to evaluate the phenological stages of two contrasting lines, B481_6 and R453, during senescence. We observed clear differences in phenological stages, confirming the results obtained in previous studies. A database with 5000 images was generated and was classified by an expert. This is important to end the subjectivity involved in decision making regarding the progression of this trait in the field and could be correlated with performance and senescence parameters that are highly associated with yield increase.

1. Introduction

Sunflower is a primary oil crop, contributing to 2% of the total harvest area worldwide and with a global gross production value of 21.4 billion USD in 2020 [1]. Despite their economic importance, sunflower breeding has shown a low increase in yield in comparison with wheat, rice, soybean, and maize [1], with its trait of early leaf senescence being one of the most relevant in terms of the gap between real and potential yields for this crop [2,3,4].
Leaf senescence is the last stage of leaf development and is characterized by an active decline in the photosynthetic rate, rupture of chloroplasts, and cell death. In annual plants, such as grain and oil crops, flowering induces senescence accompanied by nutrient remobilization (mainly nitrogen) from leaves to developing seeds [5,6,7]. The remobilization of nutrients must be coordinated with the developmental stages of the sunflower. If the senescence is anticipated to occur at flowering time or if the rate of senescence increases disproportionately during the development of the seeds, losses in yield are observed [2,4,8,9]. The plant’s ability to maintain an active photosynthetic area for longer periods might improve the recycling process and, consequently, the grain content [9].
Crop phenological information is not only important in relation to senescence, but it is also used in yield modeling and crop monitoring. The impact of climatic conditions on final crop yield depends on the phenological stage [10,11,12]. In sunflower, a reduction in the intercepted radiance by shading has a different impact on the grain number according to the developmental stage in which the shading was applied. Meanwhile, the grain set of florets in the central position on the head radius is affected when shadowing takes place in early or late post-anthesis periods, while the grain set of florets in the mid-section was only affected by early anthesis shading. Thus, the post-anthesis period is critical for determining grain number [13]. On the other hand, the reproductive period of pre-anthesis has an impact on the determination of grain size, a parameter also correlated with the maximum seed weight [14]. Finally, the early post-anthesis period is highly susceptible to intercepted radiation, being decisive for the weight and oil content of the grains [15,16].
In sunflower, the phenological stages were described by Schneiter and J.F. Miller in 1981 [17]. According to this classification, the vegetative stages are named as ‘V’ following the number of fully expanded leaves, and the reproductive stages are named from R1 (the first appearance of the inflorescence) to R9 (physiological maturity). Furthermore, stage R5 can be divided into substages R5.1 to R5.9, which describe the percentage of anthesis in the head area. The evaluation of crop phenology is conventionally identified by monitoring activities in the field. These activities provide an accurate description of crop phenology, but are also expensive and time consuming, and therefore rarely implemented on a large scale.
New technologies for describing morphophysiological parameters in sunflower fields have been developed over the past decade [18,19,20]. The use of new types of sensors in the field and their combination with mathematical models, such as deep machine learning algorithms, accelerates the evaluation time, homologates the measurement between different work groups, and improves the accuracy and resolution [21,22]. Automatic evaluation allows farmers and experts to decrease the cost of analysis by orders of magnitude, making it possible to obtain a denser temporal analysis of plant phenological stages in the field.
Neural networks (NNs) are mathematical models consisting of a series of layers formed by neurons. Briefly, we can classify NNs in three different classes: fully connected feedforward NNs (also called multi-layer perceptron (MLP)), convolutional NNs (CNNs) and recurrent NNs (RNNs). Convolutional neural networks (CNNs) have shown remarkable success in computer vision tasks such as image classification, object detection, and segmentation. CNNs are inspired by the structure of the visual cortex in animals, which is responsible for processing visual information. The key idea behind CNNs is to learn a hierarchical representation of the input image by convolving filters over the image and pooling the results to produce a feature map. This allows the network to capture local patterns in the image, such as edges and textures, and gradually build up to more abstract representations of objects and scenes. Researchers have discovered that, thanks to their capability to build a gradual understanding of objects by stacking simpler representations, convolutional neural networks are also very suitable for transfer learning. Transfer learning enables practitioners to transfer the knowledge obtained on a dataset for a specific task, to a new dataset with a possibly similar but not equivalent task.
In the context of image classification, several pretrained architectures have been introduced, mainly trained on ImageNet [23] for image classification. We will experiment with both non pretrained and pretrained architectures, including resnet18, resnet50, vgg16 and a pretrained vision transformer [24,25].Transformers are a different family of deep learning algorithms, based entirely on the attention mechanism. They have been introduced in text translation, and were shortly after adopted for other natural language processing (NLP) tasks, such as sentiment classification, topic identification [26,27,28,29] and text generation. Transformers have also been adapted in the computer vision field and applied for the first time for image classification, showing promising results.
In this study, five deep machine-learning methods, including CNNs and transformers, were evaluated for the automatic discrimination of sunflower developmental stages. The best model was then used to evaluate the progression of phenological stages in lines R453 and B481_6 on images acquired with three smartphones. These two inbred lines belonging to the INTA Sunflower Breeding Program and the INTA Manfredi Sunflower Germplasm Collection, have been described previously by the group as contrasting senescence phenotypes that also present a difference in yield.

2. Results

2.1. Dataset

A total of 5000 images were taken manually with cell phones during the growing season 2021/2022 at 14 sampling times from different angles and light conditions. The experiment was carried out with two genotypes, B481_6 and R453, with visual differences in their inflorescences. All the images were classified into five classes, as shown in Figure 1. Each class contained images of a particular developmental stage. The S1 class included all images of plants in the vegetative stages, regardless of the number of leaves. S2 included all reproductive stages described by Schneiter and J. F. Miller before the anthesis, referred to as the pre-anthesis period. S3 class is the anthesis stage and is characterized by the maturity of the ray flower, visible floret disks, stigma exposure, and dehisced anther locules. The S4 class included images belonging to the R6 reproductive stages referred to as the post-anthesis period, and no pictures of the R7 and R8 stages were taken. S5 class included images of plants at physiological maturity.

2.2. Methods and Models

Five different methods—CNN from scratch, VGG19, ResNet18, ResNet50, and VITB16—were evaluated using a balanced dataset of 450 images per class. In addition, data augmentation was implemented to augment the number of training samples artificially. Except for the CNN from scratch, transfer learning was applied.
A total of 300 epochs were performed for each method. Loss and accuracy were recorded in the validation and training sets for each epoch (Figure 2). During the training, the best model for each method was saved.
The best model for the CNN from scratch method is described in Table 1. Even so, VITB16 seems to have outperformed ResNet50 in the training set (Figure 2a,b), while, in the validation set, there are no clear differences among the Resnet18, ResNet50, VITB16, and VGG19 methods. The performance metrics of the testing set revealed notable differences among the evaluated models. The best-performing model within the ResNet50 approach exhibited remarkable accuracy (95.703%), coupled with a minimal error (0.043). Despite the ResNet50 model’s higher parameter count compared with ResNet18, it maintained a comparable prediction time, demonstrating its efficiency. VITB16 and ResNet18 models showed similar accuracy and error rates, and the performance of VGG19 and the CNN from scratch lagged, exhibiting the least favorable outcomes in terms of performance (Table 1).
When evaluating the accuracy for individual classes, the model’s notable weakness lies in distinguishing between S3 and S4. This distinction is particularly evident in the CNN from scratch and VGG19 models. Interestingly, both ResNet18 and VITB16 outperformed ResNet50 in effectively distinguishing between these two stages (Figure 3).

2.3. Sunpheno

The best model generated using the ResNet50 method was named Sunpheno. Sunpheno inferences in the test set showed that some pictures were incorrectly classified with high confidence (Supplementary Table S1). In this group of images, the largest portion belonged to the images of stage S3.
It seems that the model has problems identifying the stages in pictures taken from the top when the floret weight turns down the plant (Figure 4a), and classification of pictures taken from the side (Figure 4b,c). Finally, one should consider that the prediction can be affected when there is overlap between vegetative and reproductive organs (Figure 4d).

2.4. Case Study

To test the Supheno model, we used it to monitor the developmental stage of sunflowers in two experiments carried out at INTA Castelar during the 2020/2021 and 2021/2022 growing seasons. The pictures were taken at 12 pm with variable weather conditions. The objective of these experiments was to compare the progress of senescence of genotypes B481_6 and R453 and its correlation with plant development. These genotypes have been widely studied because of their contrasting senescence phenotypes.
In both growing seasons (Figure 5a,b), the R453 genotype exhibited an early senescence phenotype in contrast with the B481-6 genotype. R453 showed half of the total amount of leaves senescent after 1100 °Cd and a positive increase in the slope after 1300 °Cd, reducing the plant lifespan. Meanwhile, genotype B481_6 showed half of the leaves senescent at 1600 °Cd with a diminished slope after 1300 °Cd. The obtained yield, measured as a weight of 1000 g, was significantly slower in the genotype R453 in both growing seasons, as was expected for this genotype (Figure 6a,b).
Sunpheno successfully described the progression of developmental stages in both the growing seasons and genotypes. A total of 20 pictures for each sampling time and genotype were predicted (Figure 7). As time progressed, the proportion of predicted images shifted from S1 dominant at the beginning to S5 dominant at the later sampling date. No significant differences were found between genotypes, as most of the plants reached anthesis (S3) at approximately 850 °C and 950 °C. Before 168 °Cd in the growing season 21/22, Sunpheno had difficulty in assigning the label S1 to the images. This could be the result of the small size of the plants or the inference of mulch sheet.

3. Discussion

The detection of developmental stages, particularly in biological and agronomical contexts, often relies on analyzing various data points, such as images, genetic data, physiological measurements, etc. Implementing an intelligent agent (IA) for this purpose typically involves two key components, an expert eye on a field and a robust and well-supported machine-learning model [30]. A classic example is field phenotyping, which has unlocked new prospects for non-destructive field-based clean observation in plants for a large number of traits, including physiological, biotic (includes living factors such as fungi, bacteria, virus, insects, parasites, and weeds, etc.) and abiotic (includes non-living factors such as drought, flood, nutrient deficiency, senescence and other environmental factors) stress traits [31]. Determination of the phenological stages of plants is important for the growth of healthy and productive plants. The knowledge of transition times of phenological stages of a plant can provide valuable data for planning, organizing and timely execution of agricultural activities [32].
Due to the widespread availability of mobile phones and applications, tasks such as image classification can now be completed at the point of care. The deployment of biological image classification models on mobile phones can lead to several technical issues. For example, the model’s performance when trained or tested on a computer degrades significantly when the same model is deployed on a mobile phone [33].
The present study aimed to develop a tool for the automatic discrimination of sunflower phenological stages. Different deep machine-learning methods were applied to 2250 images, revealing that the ResNet50 architecture was the most accurate algorithm for this experiment, according to what happens in other systems, such as potato [34] and seldom crop systems [35].
The model generated in this study, Sunpheno, was used to evaluate changes in the developmental stages of the two genotypes with phenotypic differences in their inflorescences. These results are equivalent to those observed in previous studies [36]. R453 and B481-6 genotypes have similar phenological progressions, reaching anthesis at a similar time; however, senescence after anthesis in R453 was more pronounced than that in B481-6. This early senescence phenotype affected the observed yield of genotype R453.
Sunpheno successfully evaluated the changes in the developmental stages of the two genotypes, suggesting that it is robust enough to distinguish developmental stages, even if there are phenotypic differences between genotypes. However, additional genotypes and different growth treatments should be evaluated to confirm the application of this model.
The images in this work were taken with cell phones, which, unlike remote sensing tools, are laborious and have high dimensions as long as the image resolution level of a human is well trained for the physiological event addressed in the study, considering the elimination of the background as a key process for complex field trials and with the presence of the relevant biological information required to detect phenological stages [37,38]. However, evaluating developmental stages through a machine learning model homogenizes the results of these changes and removes subjectivity from the analyst measurements. As future work, hybrid techniques of machine vision and deep learning models can be applied in further field experiments to develop automated systems for precision agriculture based on cell phone images and a well-trained field observatory team. In addition, this model is intended as the first step in the development of a fully automatic monitoring experimental field, utilizing fixed cameras that can be installed on the ground or at the top of the antibird cages or even utilizing image acquisition from drones, as a future development for smart farming [39]. Additionally, CNN-based image analysis holds great promise for the accurate and efficient recognition of phenological stages in crops [40].
Finally, this study generated a 5000-image database, subdivided into 5 phenological stages that can be used in future computer vision studies. Hence, developing an IA machine learning model for detecting developmental stages is a multidisciplinary effort that combines data science and software engineering.

Conclusions

In this study, a novel high-precision and high-efficiency method was proposed based on the deep learning convolutional neural network (CNN) model for the automatic detection of phenological stages in sunflower. Sunpheno has been shown to be a high-throughput and low-cost method, mainly developed and evaluated using images collected in a field, and can help breeders and farmers in decision making with regard to the early leaf senescence trait in cultivated sunflower.

4. Materials and Methods

4.1. Plant Material and Experimental Conditions

Two field experiments were conducted at the INTA Castelar Experimental Station (34°60′48″ S, 58°67′33″ W) during the 2020/2021 and 2021/2022 growing seasons.
Two sunflower inbred genotypes, R453 and B481-6, from the Instituto Nacional de Tecnología Agropecuaria (INTA) Sunflower Breeding Program, previously characterized as contrasting genotypes, were sown and evaluated for the senescence phenotype [36,41]. The plants were sown at 7.2 plants/m2 and cultivated under field conditions. The plants were grown under non-limiting water conditions, and soil water was maintained by irrigation. Diseases, weeds, insects, and birds were adequately controlled. Time was expressed on a thermal time basis by daily integration of air temperature with a threshold temperature of 6 °C and with plant emergence as the thermal time origin [42].

4.2. Physiological Parameters

Three plants were tagged, and the number of senescent leaves was calculated for each sample. After visual inspection, all leaves with more than 50% chlorotic tissue were considered senescent.
The phenological stages of the plants were followed every two days (growing season 2021/2022) or four days (growing season 2020/2021) by visual inspection. A stage was determined if more than 50% of the plants belonged to that stage.
When tagged plants reached physiological maturity, their heads were harvested. Seed number and seed weight per capitulum were also measured. Yield per genotype was calculated as the weight of 1000 seeds as follows: yield (g) = (FW(g) number of seeds) × 1000 [27].

4.3. Dataset

For each plant (~200 per genotype), pictures were taken during the life cycle on 14 dates during the season 2021/2022. Pictures were taken from different angles, distances and light conditions using three different smartphones. The resolution of the cameras was 14mpx and the image sizes varied between 4656 × 3492, 3016 × 4032, and 3024 × 4032. All of the images were resized to 224 × 224.
Twenty images from each genotype and date were separated, and the experimental set growing season was 21/22. The remaining pictures were classified into 5 different development stages: S1, S2, S3, S4, and S5, containing 450 images each. This dataset was random sized and split in three balanced sets: train, validation, and test in a relation 6.1.3.
During the season 2020/2022, 20 extra pictures were taken for each genotype on six dates along the life cycle of the plants (experimental set growing season 20/21).

4.4. Methods

We tested several methods for sunflower phenological stage classification from images, including convolutional neural networks and transformer vision, highlighting the memory requirements and the processing times to better compare the networks in the context of lightweight classification.

4.4.1. Baseline CNN

We designed a CNN from scratch to serve as the baseline. The network is composed of two convolutional layers followed by a fully connected layer. The two convolutional layers apply a series of four filters each of size 3 × 3. Stride and padding are set to 1, to preserve the input width and height. Batch normalization is applied after each convolutional layer, which normalizes the output of the convolutional layer to improve the stability and speed of training. Max pooling with 2 × 2 kernel is also applied in each layer, to help the network learn features that are more robust to input translations. A ReLU activation function is then applied elementwise to the output of the normalization layer, which introduces non-linearity and helps the network learn more complex representations. After the convolutional layers, a fully connected layer is applied to project the features to the final classification scores for the five classes (Supplementary Figure S1).

4.4.2. VGG19

We chose to experiment with VGG19, a 19-layer convolutional neural network that belongs to the VGG family. This family of CNNs was introduced in [25] to perform classification on the ImageNet dataset. The authors sought explicitly to test the effect of the depth on the classification accuracy. The authors were also among the first to test very small (3 × 3) convolutional kernels in every layer. Thanks to its depth and its design, the VGG architecture achieved state of the art results in the ILSVRC-2014 competition. Specifically, VGG19 is composed of 16 convolutional layers followed by three fully connected layers. The concept is to reduce the spatial dimension of the feature maps, while increasing their number [25]. The reduction in spatial dimension is achieved using max pooling layers, while the increase in feature maps is achieved through the convolutional layers. The last three are a cascade of fully connected layers that project the features to the final output scores. Overall, VGG architectures have shown great accuracy using fewer parameters compared with counterparts such as AlexNet [25] (Supplementary Figure S2).

4.4.3. ResNet Networks (18–50)

The family of ResNet architectures was introduced in [43]. At the time, researchers noticed that very deep networks were difficult to train, and that the deeper the network, the harder it was for it to converge. ResNet is the first work to introduce the concept of skip connections. Skip connections enable the direct transfer of information from one layer to another that is not adjacent in the network architecture. One of their main benefits is that they help to mitigate the vanishing gradient problem, which is a common issue in deep neural networks. The vanishing gradient problem occurs when the gradient of the loss function becomes very small as it propagates through the network during backpropagation, making it difficult to update the weights of the earlier layers. By using skip connections, the information from the earlier layers can be directly passed to the later layers, which can help to preserve the gradient and improve the flow of information through the network. Skip connections also allow for the creation of deeper and more complex neural networks without sacrificing performance. This is because the skip connections allow for the reuse of features learned in earlier layers, which can help to improve the overall accuracy and generalization of the network. Finally, skip connections can help to improve the speed of convergence during training. By allowing the network to access information from earlier layers, skip connections can help to speed up the learning process and reduce the number of iterations required to reach convergence. We will experiment with two ResNet variants, resnet18 and resnet50. The main difference between them is the depth of the network, one is characterized by 18 layers, while the second by 50 layers. Additionally, resnet18 and resnet50 employ two different building blocks. Resnet18, given the contained number of layers, adopts a basic block formed by two 3 × 3 convolutions that reduces the spatial dimension by increasing the number of feature maps. In contrast, resnet50 adopts a basic block formed by three convolutions in series. The first exploits a 1 × 1 kernel to reduce the number of feature maps to improve the efficiency of the 3 × 3 convolution that follows. Finally, another 1 × 1 convolution brings the number of feature maps back to the initial one. This process allows the network capacity to be maintained by using fewer parameters, especially in very deep networks. In fact, this approach is also used in other ResNet variants, such as resnet-101 or resnet-152 (Supplementary Figure S3).

4.4.4. Transformer

The Transformer architecture [44], originally developed for natural language processing (NLP) tasks, has recently gained attention in image processing as an alternative to convolutional neural networks (CNNs). CNNs have been highly successful in image processing tasks, but they struggle with capturing long-range dependencies due to the limited receptive field of the convolutional kernel. The Transformer architecture, on the other hand, relies solely on self-attention mechanisms and does not rely on sequential processing. This allows it to capture dependencies between distant elements in the input sequence without being limited by the receptive field of the convolutional kernel. Moreover, the self-attention mechanism of the Transformer architecture enables it to attend to different parts of the input image, making it well-suited for image processing tasks that require the capturing of relationships between non-adjacent pixels. Recent studies have shown that the Transformer architecture can achieve competitive results on several tasks, such as image classification, object detection, and segmentation. Its efficient use of memory and its ability to handle long-range dependencies make it a particularly appealing choice for these tasks, especially in cases where the input image has complex and non-local dependencies. While CNNs remain the dominant approach for image processing tasks, the Transformer architecture presents a promising alternative, with its ability to capture non-local dependencies and attend to different parts of the input image; as a result, we decided to include it in our analysis (Supplementary Figure S4).

4.5. Transfer Learning

Transfer learning is a machine-learning technique in which a model trained on one task is used as a starting point for a different but related task. The so-called “pre-trained” model is then used to extract general features that can be used as input for a new model trained on a specific task. In the computer vision community, models pre-trained on the ImageNet [24] dataset have been a longstanding starting point for a series of downstream tasks, including image classification.
One of the key benefits of transfer learning is that it can significantly reduce the amount of data required to train a new model. By using a pre-trained model as a starting point, the knowledge already captured in the pre-trained weights can be leveraged to improve performance and achieve a faster convergence. Additionally, transfer learning can help mitigate the problem of overfitting, as the pre-trained model has already learned general features that can be used to regularize the new model. In the context of image classification, transfer learning is usually achieved by removing the last classification head and replacing it with a different head to accommodate for the number of output classes. The network is then trained by freezing the parameters of all but the last classification head. This prevents the pretrained weights from changing during training and allows the classification head to adapt to the given task and the extracted features.

4.6. Data Augmentation

Data augmentation is a technique used to artificially increase the size of a dataset by creating new, modified versions of existing data. With classification, several different transformations can be applied to augment the dataset. In this work we applied transformation to the input images to increase the dataset variability. Randomly, 20% of the input images are rotated either in 90, 180, or 270 degrees; 20% are changed by increasing or decreasing the brightness; and 20% are horizontally or vertically flipped.

4.7. Experimental Operation Environment

The proposed model was implemented in Python using the PyTorch library and trained with an NVIDIA Quadro P2000 GPU (5 GB). The initial learning rate is 0.007, which varies in a Poly manner. The maximum number of epochs used for training is 300, while the batch size is 8. We used an Adam optimizer with a weight decay factor of 0.0001 on the parameters to prevent overfitting.

4.8. Performance Evaluation

Precision, recall, and macro F1 score are important metrics used to evaluate the performance of classification models. Precision (Pr) measures the proportion of correctly predicted positive instances among all instances predicted as positive, highlighting the model’s accuracy in identifying true positives (TP) while minimizing false positives (FP). Recall (Rc), on the other hand, assesses the ratio of correctly predicted positive instances to the total actual positive instances, focusing on the model’s ability to capture all positive instances and reducing false negatives. Macro F1 score combines both precision and recall, providing a balanced measure of a model’s effectiveness by considering their harmonic mean. It is particularly useful when dealing with imbalanced classes, as it gives equal weight to each class and offers a comprehensive assessment of a model’s performance across all classes. The overall performance of each method was evaluated with the following metrics, as follows:
P r e c i s i o n = P r i = T P i T P i + F P i
R e c a l l = R c i = T P i T P i + F N i
M a r c o   A v e r a g e   P r e c i s i o n = M P r = i P r i N
M a r c o   A v e r a g e   R e c a l l = M R c = i R c i N
M a r c o   F 1   s c o r e = 2 × M R c × M P r M R c + M P r
E r r o r = 1 M R c
C o n f i d e n c e   i n t e r v a l = 1.96 × M R c × E r r o r N s 2
A c c u r a c y = M a r c o   A v e r a g e   R e c a l l × 100
A c c u r a c y ± = C o n f i d e n c e   i n t e r v a l × 100
where i denotes the class (S1, S2, S3, S4, S5).

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/plants13141998/s1.

Author Contributions

S.A.B.L. and P.F. designed the field assay; S.A.B.L. and M.A.C. performed the experiments. R.R. and G.H. performed data analysis, wrote open codes and executed routines. S.A.B.L., P.F., M.A.C. and F.M. conceived and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by INTA PE 1131022, 1131043; ANPCyT Préstamo BID PICT 2019–01602 and PIP CONICET PIP 11220120100262CO. The funding bodies provided financial support and were not involved in the design of the study, collection, interpretation or analysis of data and in writing the manuscript.

Data Availability Statement

For making inferences in images a google share folder was created (https://drive.google.com/drive/folders/1q39dsbNmvoh8t_jSCSsprHNWZw78yYYV?usp=drive_link (accessed on 24 June 2024)). The subfolder Using_the_model contains Google collabs (or Python notebook) that can be used as a tutorial. The Google collabs connects to the Sunpheno model in order to make inferences about the images save on the subfolder ‘pics’. After running the notebook, the output table with the name of the images, inference value, and confidence in the prediction are saved in the subfolder ‘results’. The whole project containing the scrips used are available in the same google share folder ‘Sunflower_Stages’. Image dataset supporting the findings of this study are available upon reasonable request, due to the storage size required. A public repository hosting the source codes and a sample subset of the dataset used in the current work will be made promptly available at INTA hosting.

Acknowledgments

We express our gratitude to the Organización Italo Latino Americana (IILA) for their generous financial support, which has been integral to the success of our bioinformatics research project. Their commitment to advancing scientific endeavors has significantly contributed to the achievements of our work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. FAOSTAT. Library of Congress: Washington, DC, USA. Available online: https://lccn.loc.gov/2005617801 (accessed on 25 June 2024).
  2. Sadras, V.O.; Echarte, L.; Andrade, F.H. Profiles of leaf senescence during reproductive growth of sunflower and maize. Ann. Bot. 2000, 85, 187–195. [Google Scholar] [CrossRef]
  3. Dosio, G.A.A.; Aguirrezábal, L.A.N. Variaciones del rendimiento en girasol. Identificando las causas. Rev. Agromercado Cuad. Girasol 2004, 90, 7–10. [Google Scholar]
  4. De la Vega, A.J.; Cantore, M.A.; Sposaro, M.M.; Trápani, N.; Pereira, M.L.; Hall, A.J. Canopy stay-green and yield in non-stressed sunflower. Field Crops Res. 2011, 121, 175–185. [Google Scholar] [CrossRef]
  5. Himelblau, E.; Amasino, R.M. Nutrients mobilized from leaves of Arabidopsis thaliana during leaf senescence. J. Plant Physiol. 2001, 158, 1317–1323. [Google Scholar] [CrossRef]
  6. Agüera, E.; Cabello, P.; de la Haba, P. Induction of leaf senescence by low nitrogen nutrition in sunflower (Helianthus annuus) plants. Physiol. Plant. 2010, 138, 256–267. [Google Scholar] [CrossRef] [PubMed]
  7. Moschen, S.; Gialdi, A.I.L.; Paniego, N.; Fernandez, P.; Heinz, R.A. Sunflower Leaf Senescence: A Complex Genetic Process with Economic Impact on Crop Production; Dorszewska, J., Kozubski, W., Eds.; InTech Open: London, UK, 2017. [Google Scholar] [CrossRef]
  8. Gregersen, P.L.; Culetic, A.; Boschian, L.; Krupinska, K. Plant senescence and crop productivity. Plant Mol. Biol. 2013, 82, 603–622. [Google Scholar] [CrossRef] [PubMed]
  9. Kusaba, M.; Tanaka, A.; Tanaka, R. Stay-green plants: What do they tell us about the molecular mechanism of leaf senescence. Photosynth. Res. 2013, 117, 221–234. [Google Scholar] [CrossRef] [PubMed]
  10. Sakamoto, T.; Gitelson, A.A.; Arkebauer, T.J. MODIS-based corn grain yield estimation model incorporating crop phenology information. Remote Sens. Environ. 2013, 131, 215–231. [Google Scholar] [CrossRef]
  11. Observed Northward Migration of Agro-Climate Zones in Europe Will Further Accelerate Under Climate Change—Ceglar—2019—Earth’s Future—Wiley Online Library. Available online: https://agupubs.onlinelibrary.wiley.com/doi/10.1029/2019EF001178 (accessed on 24 June 2024).
  12. Du, Q.; Liu, H.; Li, Y.; Xu, L.; Diloksumpun, S. The effect of phenology on the carbon exchange process in grassland and maize cropland ecosystems across a semiarid area of China. Sci. Total Environ. 2019, 695, 133868. [Google Scholar] [CrossRef]
  13. Cantagallo, J.E.; Medan, D.; Hall, A.J. Grain number in sunflower as affected by shading during floret growth, anthesis and grain setting. Field Crops Res. 2004, 85, 191–202. [Google Scholar] [CrossRef]
  14. Castillo, F.M.; Canales, J.; Claude, A.; Calderini, D.F. Expansin genes expression in growing ovaries and grains of sunflower are tissue-specific and associate with final grain weight. BMC Plant Biol. 2018, 18, 327. [Google Scholar] [CrossRef] [PubMed]
  15. Aguirreza, L.A.N.; Lavaud, Y.; Dosio, G.A.A.; Izquierdo, N.G.; Andrade, F.H.; González, L.M. Intercepted Solar Radiation during Seed Filling Determines Sunflower Weight per Seed and Oil Concentration. Crop Sci. 2003, 161, 152–161. [Google Scholar] [CrossRef]
  16. Dosio, G.A.A.; Izquierdo, N.G.; Bordoy, E.I.N.; Aguirrezábal, L.A.N. Leaf senescence did not account for variations in grain and oil yield observed in sunflower under radiation limiting conditions. Agric. For. Meteorol. 2020, 291, 108032. [Google Scholar] [CrossRef]
  17. Schneiter, A.A.; Miller, J.F. Description of sunflower growth stages. Crop Sci. 1981, 21, 901–903. [Google Scholar] [CrossRef]
  18. Gosseau, F.; Blanchet, N.; Varès, D.; Burger, P.; Campergue, D.; Colombet, C.; Gody, L.; Liévin, J.-F.; Mangin, B.; Tison, G.; et al. Heliaphen, an outdoor high-throughput phenotyping platform for genetic studies and crop modeling. Front. Plant Sci. 2019, 9, 1908. [Google Scholar] [CrossRef] [PubMed]
  19. Malik, A.; Vaidya, G.; Jagota, V.; Eswaran, S.; Sirohi, A.; Batra, I.; Rakhra, M.; Asenso, E. Design and Evaluation of a Hybrid Technique for Detecting Sunflower Leaf Disease Using Deep Learning Approach. J. Food Qual. 2022, 2022, 9211700. [Google Scholar] [CrossRef]
  20. Pérez-Ortiz, M.; Peña, J.M.; Gutiérrez, P.A.; Torres-Sánchez, J.; Hervás-Martínez, C.; López-Granados, F. A semi-supervised system for weed mapping in sunflower crops using unmanned aerial vehicles and a crop row detection method. Appl. Soft Comput. 2015, 37, 533–544. [Google Scholar] [CrossRef]
  21. de Luna, R.G.; Dadios, E.P.; Bandala, A.A.; Vicerra, R.R.P. Tomato Growth Stage Monitoring for Smart Farm Using Deep Transfer Learning with Machine Learning-based Maturity Grading. AGRIVITA J. Agric. Sci. 2020, 42, 24–36. [Google Scholar] [CrossRef]
  22. Computer Vision-Based Phenotyping for Improvement of Plant Productivity: A Machine Learning Perspective|GigaScience|Oxford Academic. Available online: https://academic.oup.com/gigascience/article/8/1/giy153/5232233?login=false (accessed on 24 June 2024).
  23. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021. [Google Scholar] [CrossRef]
  25. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015. [Google Scholar] [CrossRef]
  26. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2019. [Google Scholar] [CrossRef]
  27. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  28. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. arXiv 2020. [Google Scholar] [CrossRef]
  29. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv 2023. [Google Scholar] [CrossRef]
  30. Singh, A.; Ganapathysubramanian, B.; Singh, A.K.; Sarkar, S. Machine Learning for High-Throughput Stress Phenotyping in Plants. Trends Plant Sci. 2016, 21, 110–124. [Google Scholar] [CrossRef] [PubMed]
  31. Deery, D.; Jimenez-Berni, J.; Jones, H.; Sirault, X.; Furbank, R. Proximal remote sensing buggies and potential applications for field-based phenotyping. Agronomy 2014, 4, 349–379. [Google Scholar] [CrossRef]
  32. Gulac, F.; Bayazit, U. Plant and Phenology Recognition from Field Images Using Texture and Color Features. In Proceedings of the 2018 Innovations in Intelligent Systems and Applications (INISTA), Thessaloniki, Greece, 3–5 July 2018; pp. 1–6. [Google Scholar] [CrossRef]
  33. Muneeb, M.; Feng, S.F.; Henschel, A. Deep learning pipeline for image classification on mobile phones. In Proceedings of the 9th International Conference on Artificial Intelligence and Applications (AIAPP 2022), Vancouver, BC, Canada, 28–29 May 2022; pp. 1–20. [Google Scholar] [CrossRef]
  34. Shaheed, K.; Qureshi, I.; Abbas, F.; Jabbar, S.; Abbas, Q.; Ahmad, H.; Sajid, M.Z. EfficientRMT-Net—An Efficient ResNet-50 and Vision Transformers Approach for Classifying Potato Plant Leaf Diseases. Sensors 2023, 23, 9516. [Google Scholar] [CrossRef]
  35. Bayazit, U.; Altilar, D.; Bayazit, N. Classification and phenological staging of crops from in situ image sequences by deep learning. Turk. J. Electr. Eng. Comput. Sci. 2022, 30, 1299–1316. [Google Scholar] [CrossRef]
  36. Gialdi, A.I.L.; Moschen, S.; Villán, C.S.; Fernández, M.P.L.; Maldonado, S.; Paniego, N.; Heinz, R.A.; Fernandez, P. Identification and characterization of contrasting sunflower genotypes to early leaf senescence process combining molecular and physiological studies (Helianthus annuus L.). Plant Sci. 2016, 250, 40–50. [Google Scholar] [CrossRef]
  37. Ngugi, L.C.; Abdelwahab, M.; Abo-Zahhad, M. Tomato leaf segmentation algorithms for mobile phone applications using deep learning. Comput. Electron. Agric. 2020, 178, 105788. [Google Scholar] [CrossRef]
  38. Restrepo-Arias, J.F.; Branch-Bedoya, J.W.; Awad, G. Image classification on smart agriculture platforms: Systematic literature review. Artif. Intell. Agric. 2024, 13, 1–17. [Google Scholar] [CrossRef]
  39. Darwin, B.; Dharmaraj, P.; Prince, S.; Popescu, D.E.; Hemanth, D.J. Recognition of Bloom/Yield in Crop Images Using Deep Learning Models for Smart Agriculture: A Review. Agronomy 2021, 11, 646. [Google Scholar] [CrossRef]
  40. Deng, R.; Jiang, Y.; Tao, M.; Huang, X.; Bangura, K.; Liu, C.; Lin, J.; Qi, L. Deep learning-based automatic detection of productive tillers in rice. Comput. Electron. Agric. 2020, 177, 105703. [Google Scholar] [CrossRef]
  41. Moschen, S.; Marino, J.; Nicosia, S.; Higgins, J.; Alseekh, S.; Astigueta, F.; Luoni, S.B.; Rivarola, M.; Fernie, A.R.; Blanchet, N.; et al. Exploring gene networks in two sunflower lines with contrasting leaf senescence phenotype using a system biology approach. BMC Plant Biol. 2019, 19, 446. [Google Scholar] [CrossRef]
  42. Kiniry, J.R.; Blanchet, R.; Williams, J.R.; Texier, V.; Jones, K.; Cabelguenne, M. Sunflower simulation using the EPIC and ALMANAC models. Field Crops Res. 1992, 30, 403–423. [Google Scholar] [CrossRef]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015. [Google Scholar] [CrossRef]
  44. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
Figure 1. Image dataset classified into five classes based on Schneiter and J. F. Miller phenological stages description.
Figure 1. Image dataset classified into five classes based on Schneiter and J. F. Miller phenological stages description.
Plants 13 01998 g001
Figure 2. Loss and accuracy during training were evaluated in the training and validation sets in each epoch. (a) Training loss, (b) training accuracy, (c) validation loss, and (d) validation accuracy. The first 100 epochs are shown.
Figure 2. Loss and accuracy during training were evaluated in the training and validation sets in each epoch. (a) Training loss, (b) training accuracy, (c) validation loss, and (d) validation accuracy. The first 100 epochs are shown.
Plants 13 01998 g002aPlants 13 01998 g002b
Figure 3. Confusion matrices for the sunflower datasets obtained in the test set. (a) CNN from scratch, (b) VGG19, (c) ResNet18, (d) ResNet50, and (e) VITB16. For each matrix, the rows show the true development stage, whereas the columns show the predictions from the CNN architecture. The accuracy of the prediction of each class is shown. Accuracy values from 0 to 1 are represented from dark purple to light yellow.
Figure 3. Confusion matrices for the sunflower datasets obtained in the test set. (a) CNN from scratch, (b) VGG19, (c) ResNet18, (d) ResNet50, and (e) VITB16. For each matrix, the rows show the true development stage, whereas the columns show the predictions from the CNN architecture. The accuracy of the prediction of each class is shown. Accuracy values from 0 to 1 are represented from dark purple to light yellow.
Plants 13 01998 g003
Figure 4. Examples of images incorrectly classified by the best model, ResNet50, with high confidence in the prediction. (a) Image of phenological stage S3, incorrectly classified as S4 with a confidence of 0.74. (b) Image of phenological stage S4, incorrectly classified as S2 with a confidence of 0.77. (c) Image of phenological stage S3, incorrectly classified as S4 with a confidence of 0.79. (d) Image of phenological stage S2, incorrectly classified as S1, with a confidence of 0.65.
Figure 4. Examples of images incorrectly classified by the best model, ResNet50, with high confidence in the prediction. (a) Image of phenological stage S3, incorrectly classified as S4 with a confidence of 0.74. (b) Image of phenological stage S4, incorrectly classified as S2 with a confidence of 0.77. (c) Image of phenological stage S3, incorrectly classified as S4 with a confidence of 0.79. (d) Image of phenological stage S2, incorrectly classified as S1, with a confidence of 0.65.
Plants 13 01998 g004
Figure 5. Number of senescent leaves over time. (a) Growing Season 20/21, with the measurements taken on five sampling dates. (b) Growing Season 21/22, with the measurements taken on fourteen sampling dates. The time was recorded as the thermal time after emergence (°Cd). The red and blue lines represent the genotypes B481_6 and R453, respectively.
Figure 5. Number of senescent leaves over time. (a) Growing Season 20/21, with the measurements taken on five sampling dates. (b) Growing Season 21/22, with the measurements taken on fourteen sampling dates. The time was recorded as the thermal time after emergence (°Cd). The red and blue lines represent the genotypes B481_6 and R453, respectively.
Plants 13 01998 g005
Figure 6. Box plot showing the weight in grams of 1000 seeds for R453 and B481_6 genotypes. (a) Growing Season 20/21 (b) Growing Season 21/22. Outliers are shown as open circles outside the boxes. The top and bottom of the box represent the first quartile (Q1) and third quartile (Q3), respectively. This range is known as the interquartile range (IQR) and contains the middle 50% of the data. A green line inside the box represents the median (Q2) of the data.
Figure 6. Box plot showing the weight in grams of 1000 seeds for R453 and B481_6 genotypes. (a) Growing Season 20/21 (b) Growing Season 21/22. Outliers are shown as open circles outside the boxes. The top and bottom of the box represent the first quartile (Q1) and third quartile (Q3), respectively. This range is known as the interquartile range (IQR) and contains the middle 50% of the data. A green line inside the box represents the median (Q2) of the data.
Plants 13 01998 g006
Figure 7. Sunflower stages predicted with Sunpheno. (a) Genotype R453, growing season 20/21. (b) Genotype R481-6, growing season 20/21. (c) Genotype R453, growing season 21/22. (d) Genotype R481-6, growing season 21/22. The time was recorded as the thermal time after emergence (°Cd). For each time, 20 images were predicted. Yellow bars indicate stage S1, green S2, purple S3, red S4 and blue S5.
Figure 7. Sunflower stages predicted with Sunpheno. (a) Genotype R453, growing season 20/21. (b) Genotype R481-6, growing season 20/21. (c) Genotype R453, growing season 21/22. (d) Genotype R481-6, growing season 21/22. The time was recorded as the thermal time after emergence (°Cd). For each time, 20 images were predicted. Yellow bars indicate stage S1, green S2, purple S3, red S4 and blue S5.
Plants 13 01998 g007
Table 1. Performance metrics.
Table 1. Performance metrics.
CNN from ScratchVGG19ResNet18ResNet50VITB16
Macro average precision 0.830.910.950.960.94
Macro average recall 0.840.910.950.960.94
Macro F1 score 0.840.910.950.960.94
Error 0.160.090.050.040.06
Confidence interval (95%, z = 1.96) 0.030.020.020.020.02
Accuracy 83.7090.9694.5295.7094.22
Accuracy +/−2.792.161.721.531.76
Number of parameters63,001139,601,73311,179,07723,518,27785,802,501
Time of test 675 images (s)63,09118,55217,25218,54218,049
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luoni, S.A.B.; Ricci, R.; Corzo, M.A.; Hoxha, G.; Melgani, F.; Fernandez, P. Sunpheno: A Deep Neural Network for Phenological Classification of Sunflower Images. Plants 2024, 13, 1998. https://doi.org/10.3390/plants13141998

AMA Style

Luoni SAB, Ricci R, Corzo MA, Hoxha G, Melgani F, Fernandez P. Sunpheno: A Deep Neural Network for Phenological Classification of Sunflower Images. Plants. 2024; 13(14):1998. https://doi.org/10.3390/plants13141998

Chicago/Turabian Style

Luoni, Sofia A. Bengoa, Riccardo Ricci, Melanie A. Corzo, Genc Hoxha, Farid Melgani, and Paula Fernandez. 2024. "Sunpheno: A Deep Neural Network for Phenological Classification of Sunflower Images" Plants 13, no. 14: 1998. https://doi.org/10.3390/plants13141998

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop