Next Article in Journal
Growth, Biochemical Traits, Antioxidant Enzymes, and Essential Oils of Four Aromatic and Medicinal Plants Cultivated in Phosphate-Mine Residues
Previous Article in Journal
Water-Light Interaction and Its Effect on the Morphophysiology of Cedrela fissilis Vell. Seedlings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Stream Architecture Enhanced by Soft-Attention Mechanism for Plant Species Classification

Mixed Reality and Interaction Laboratory, Department of Software, Sejong University, Seoul 05006, Republic of Korea
*
Author to whom correspondence should be addressed.
Plants 2024, 13(18), 2655; https://doi.org/10.3390/plants13182655
Submission received: 12 August 2024 / Revised: 14 September 2024 / Accepted: 18 September 2024 / Published: 22 September 2024
(This article belongs to the Section Plant Modeling)

Abstract

:
Plants play a vital role in numerous domains, including medicine, agriculture, and environmental balance. Furthermore, they contribute to the production of oxygen and the retention of carbon dioxide, both of which are necessary for living beings. Numerous researchers have conducted thorough research in the classification of plant species where certain studies have focused on limited numbers of classes, while others have employed conventional machine-learning and deep-learning models to classify them. To address these limitations, this paper introduces a novel dual-stream neural architecture embedded with a soft-attention mechanism specifically developed for accurately classifying plant species. The proposed model utilizes residual and inception blocks enhanced with dilated convolutional layers for acquiring both local and global information. Following the extraction of features, both streams are combined, and a soft-attention technique is used to improve the distinct characteristics. The efficacy of the model is shown via extensive experimentation on varied datasets, including several plant species. Moreover, we have contributed a novel dataset that comprises 48 classes of different plant species. The results demonstrate a higher level of performance when compared to current models, emphasizing the capability of the dual-stream design in improving accuracy and model generalization. The integration of a dual-stream architecture, dilated convolutions, and soft attention provides a strong and reliable foundation for the botanical community, supporting advancement in the field of plant species classification.

1. Introduction

Plants play a crucial part in people’s lives for a variety of reasons, including growing concerns about a deficiency of nutritious food around the globe with an exponentially growing population and changing climates globally [1]. Modern plant agriculture techniques are enormous because they have a substantial impact on the national economy and personal aspects of people’s lives [2]. Plant species are useful for medicine, food, and industrial applications. They also play a vital role in a country’s economic growth. Environmentalists are also interested in the automation of plant identification by developing a system based on spatiotemporal characteristics extracted from their structure and appearance. Conventional plant species recognition methods are costly and time-consuming because they require the manual involvement of human specialists [3]. Some plants are on the verge of extinction. As a result, it is critical to establish a database for plant protection. The most basic or traditional procedure is the naked-eye examination of the plant, and this procedure entails constant expert observation of a variety of farm areas [4,5], which is a lengthy and costly procedure. The automatic classification of the plant species is key to achieving plant protection and safety and making the process easy for classifying the plant species [6]. Plant leaves are generally the most prominent parts for identifying species, as opposed to flowers, which are only available for a limited time. As a result, the leaves are an optimal choice for developing a system for the automation of plant recognition. The leaves play an important role in understanding plant genetic relationships, structure, and plant development. Plant identification is a difficult task, including for botanists, due to the existence of a large number of species in the world [7,8]. Botanists use leaf recognition technology to classify the species of various plants. Plants have distinct characteristics that vary in many ways, such as size, texture, color, and shape, and they are different due to these characteristics [9]. There are several “Computer-Aided Detection” (CAD) techniques that have been used for plant recognition based on their leaves, where in recent years, [10,11] achieved significant classification accuracy in plant species recognition.
Machine-learning and deep-learning approaches have become more popular for classifying plant species, providing potential opportunities for automating plant species identification [12]. Several researchers have investigated the use of different machine-learning algorithms and deep-learning architectures for this problem, including both conventional classifiers and advanced neural networks [13,14]. Although these systems failed to obtain high levels of accuracy, there is a prevailing tendency to choose simpler structures since they are easier to execute and deploy to the edge devices. Fundamental designs often consist of shallow neural networks or traditional machine-learning methods like Support Vector Machines (SVMs) or Random Forests. Although these approaches may provide acceptable outcomes in certain circumstances, they often demonstrate constraints when confronted with complicated datasets or a wide range of plant species. Deep-learning models, which are renowned for their capacity to acquire intricate hierarchical representations from unprocessed data, have gained significant popularity in the field of plant species categorization [12]. Nevertheless, several deep-learning models used in this particular domain encounter drawbacks such as overfitting, computational complexity, and lack of interpretability. Despite these difficulties, researchers persist in investigating innovative methods and structures to surmount these constraints and improve the precision and resilience of plant species classification.
Plant species recognition and classification accuracy are essential in many fields, including environmental protection and agriculture. Making use of recent developments in deep learning, a unique model architecture has been created explicitly for the accurate visual cue-based recognition of plant species. Our proposed cutting-edge model combines a two-stream architecture, using RGB image data for both streams as an input to enhance its discriminative power. Dilated Convolutional Neural Network layers (DiCNN) are utilized in the first streams to analyze RGB images and extract complex spatial hierarchies and patterns. In addition, a second stream processes the same data and acquires different types of patterns and features from the input data. Additionally, a soft-attention mechanism is employed in the model, which focuses on the most prominent features and ignores the less prominent features. This attention module enhanced the model’s capability to deal with complex images and made it more expert in identifying the different plant species. By using the attention process and encapsulating the dual-stream integration, our proposed model has the potential to correctly classify the plant species from their visual structures. This all-encompassing architecture, which combines several dilated convolutional, pooling, residual, and inception layers, creates a strong foundation for classifying plant species. It uses a variety of techniques to identify complex features that are necessary for the precise identification of species. The proposed technique presents the following contribution.
  • We contribute a novel dataset of indoor plant species, which comprises 48 different classes. This dataset is a valuable resource for further progressing plant species recognition with diverse classes, and it allows researchers to further develop more effective techniques in the field of plant species classification.
  • We developed a dual-stream network that utilizes two parallel streams, such as the residual block and inception block, embedded with dilated convolutional layers to extract more prominent features from both streams. In contrast to existing pre-trained models, that used shallow architecture, our proposed model used two parallel streams to enhance the feature-extraction process across different plant species for accurate classification.
  • Our proposed dual-stream model is further strengthened by using a soft-attention mechanism. The integration of a soft-attention mechanism enhances the model’s adaptability to focus on more prominent features by dynamically amplifying and weighting salient data from both streams while ignoring the irrelevant features.
  • We performed a detailed ablation study and comparative analysis of the proposed model with other baseline models. The results indicate that the proposed model outperforms all the other baseline models in terms of accuracy, precision, recall, and F1 score.
The subsequent sections of the article are structured as follows: Section 2 discusses the theoretical background, major trends, and substantial drawbacks associated with existing algorithms. Section 3 provides a detailed overview of each module employed in the proposed network. Additionally, Section 4 includes the experimental setting, dataset description, and comparative analysis in terms of quantitative and qualitative against various competitive benchmarks, followed by a detailed ablation study that represents the effect of each module used in the proposed method. Finally, we conclude the article and provide the future direction of the research in Section 5.

2. Literature Review

In recent times, computer vision research has found interest in the study of plant classification through image processing [15]. The organs of plants, such as their leaves, roots, fruits, flowers, skins, stems, and seeds, can be used to categorize and identify different species of plants. Compared to other organs, leaves are essentially in a plane state and have a stable shape and structure, which is advantageous for two-dimensional image processing. Leaves are abundant and easily harvested throughout practically all growth seasons, whereas roots are inaccessible and flowers are only available during flowering seasons. That is the reason plant species classification systems most typically depend on leaves. Since leaf-based plant species recognition does not require the expert knowledge of a botanist, it is preferred over molecular biology techniques [16,17]. Plant species can be identified by the form, color, texture, and other properties of their leaves. However, the leaf color trait is rarely utilized to identify plant species as leaf colors change with the seasons and the environment. In nature, the leaf-form properties are relatively stable. In morphology, the diversity of leaf forms is a crucial foundation for plant classification and recognition [18].
Researchers have recently combined neural networks and computation devices to address many problems in this modern world, specifically in providing assistance to human beings. They have indicated that image-based assessment systems offer more accurate and consistent findings than human visual inspections [19] in plant species classification and plant disease detection [20,21]. Accurate classification of medicinal plants is vital for individuals involved in the preparation of Ayurvedic medicine, botanists, forest department offices, farmers, and practitioners. There has been plenty of effort into classifying objects using various methodologies, where [22] established CNN as a fundamental deep-learning tool for introducing deep-learning model strategies in classification and detection. Deep-learning models have been employed in agriculture to a limited level in recent years. One kind of dynamic model that can be useful for classification applications is CNNs. There are numerous CNN models for classification, including GoogLeNet [23], AlexNet [24], ResNet50, ResNet18, ResNet101 [25], DenseNet [26], VGG16, VGG19 [27], SqueezeNet [28], and various others. Using AlexNet and GoogLeNet, Mohanty et al. [29] were able to classify 14 distinct plant leaves with an accuracy of 99.27% and 99.34%. The authors have used a variety of input data, including color images, segmented images, and grayscale images. Another research [30] used a CNN model to classify plant leaf data and obtained 86.2% accuracy. Some studies focused on developing comprehensive systems for plant classification, wherein various machine-learning methods and feature-extraction techniques were employed [31]. These studies collectively illustrated the efficacy of integrating advanced imaging techniques with machine-learning technologies for plant species classification, which remained a critical aspect of research in plant science and phytotechnology. Likewise, the researchers in [32] presented CNN-based plant classification (CNN-PC) models for efficient plant species identification. The model generated pre-trained weights based on input plant image data and handled dependencies. The model’s performance was evaluated using the Vietnam Plant (VNP-200) dataset, which achieved a classification accuracy of 96.42% across all 200 classes. Moreover, the authors in [33] proposed a novel feature-extraction method using semi-supervised learning techniques and masked autoencoder architecture, MAE SGD, which achieved nearly 94% accuracy in classifying 710 plant species on the QuangNamForestPlant dataset. In the efforts to classify plant leaves with high accuracy, Barre et al. [34] used the LeafSnap, Foliage, and Flavia datasets for the classification of different classes using their suggested model LeafNet, where 60 classes from the Foliage dataset and 184 classes from LeafSnap were classified with an accuracy of 95.8% and 86.3%, respectively, while on 32 classes in the Flavia dataset, they obtained 97.9% accuracy. A deep CNN model in [35] used the “Multilayer Perceptron” (MLP) classifier, which achieved an accuracy of 97.7% on 44 classes in the MalayaKew dataset, and this accuracy was increased to 98.1% with the SVM classifier. Additionally, a study [36] conducted a plant classification task that utilized geometric characteristics in preliminary processing. They presented the 3SN Siamese network, which gained knowledge for the leaf classification task from both spatial and structural characteristics. It successfully classified 10 plant species from the Fla-via dataset with an accuracy of 90%. Gao et al. [37] completed a LifeCLEF Plant Identification task with an accuracy of 84.2%. Similarly, Zhu et al. [38] used a two-way-attention CNN model to recognize plant families and further identify plant classes for the four different datasets. In [39], the accuracy of classifying medicinal plants using the AlexNet model was 94.87%, while the Ayurleaf CNN model achieved 95.06% accuracy for 20 species of self-collected medicinal plant data.
Duong-Trung et al. [40] used the MobileNet model to achieve 98.5% classification accuracy. Liu et al. [41] developed a 10-layer CNN model for plant leaf classification and obtained an accuracy of 87.92% across 32 classes. With the LeafSnap dataset, the researchers in [11] implemented a ResNet model that achieved a classification accuracy of 93.09%. The authors in [7] completed plant leaf classification on images taken by [42] with the help of an Apple iPad device where the Deep Neural Network (DNN) model achieved an accuracy of 91.17%, while the CNN model has 95.58% accuracy. Using photos taken with mobile devices, Yang et al. [8] performed leaf classification of various plants with complicated backgrounds. The classification accuracy for the Inception ResNetV2 model, VGG16, and VGG 19 model was 89.6%, 91.5%, and 92.4%, respectively. Additionally, a study [43] that employed the AlexNet model to identify berry plants was able to achieve 97.80% accuracy for the three classes of data that the researchers collected themselves. The authors of [44] presented a novel method for classifying plants that combined key point identification with SIFT, morphological transformations, and leaf edge features. With a 95.62% accuracy rate on the PlantVillage dataset, the method outperformed conventional techniques that classified plant species based only on leaf shape and texture. Moreover, the importance of deep learning is demonstrated by a study [45] that used transfer learning with VGG-19 for plant recognition in Swedish leaves and achieved an amazing accuracy of 99.70% in identifying 15 different tree classes. Furthermore, using leaflet features to classify plant species, the paper [46] presented the Conv2D Xception architecture with the Adadelta Gradient Descent (CXAGD) deep-learning model. Compared to other CNN models, it achieved high accuracy (97.85%), precision (97.42%), recall (97.75%), and F-score (97.76%). Using the CXAGD, the model was trained on a dataset of 4500 vegetable leaflets. Using Canny edge detection and VGG16 on the Flavia dataset with 15 classes, the authors of [47] used leaf vein patterns to identify plant species with 95% accuracy. The approach highlighted how crucial it is to identify unique qualitative traits from leaf pictures to accurately classify plants. A concise description of various benchmarks over distinct datasets, along with their techniques, is shown in Table 1. In addition, the general limitations associated with the existing techniques are provided in the following points:
  • The existing techniques often struggle with generalization to diverse environmental conditions and class imbalances, necessitating more robust algorithms.
  • Various existing algorithms of both machine learning and deep learning require significant computational resources for training and deployment, posing challenges for resource-constrained environments.
  • The publicly available datasets contain diversity and class imbalance problems, which hinder the model training and generalization, highlighting the need for improved dataset curation and augmentation approaches.

3. Proposed Methodology

This section explains the proposed methodology of our technique, where we have explained the details of our proposed dual-stream architecture embedded with a soft-attention mechanism. We performed an ablation study that involved various methods to improve our model accuracy in terms of plant species classification. Firstly, we have used a hybrid model with different residual and inception blocks for classification. Secondly, we used these blocks, followed by an attention mechanism, which improves the model performance. An attention mechanism is a fundamental neural network approach for classification tasks that allows neural networks to selectively highlight pertinent features or portions of the input data. Thirdly, we have used the residual block and inspection block in the dual stream, where the residual block is in a separate stream and the inception block is used in a separate stream for the extraction of pertinent features from both streams. Lastly, we proposed our more accurate and efficient model, where we use a dual-stream network of residual block and inception block and then concatenate them and pass them to the soft-attention mechanism. Adding soft-attention mechanisms enhances the neural network’s ability to recognize and utilize important information, which results in increased classification accuracy and resilience over a variety of datasets and classes.

3.1. Proposed Dual-Stream Approach

Our proposed approach presents a novel dual-stream architecture that utilizes the two parallel blocks enhanced with dilated convolutional layers, which is skilled in capturing the most prominent features from the input data. By using the shortcut connections, residual blocks facilitate feature learning and gradient flow efficiently, and these connections also help in vanishing gradient problems, which is common in deep-learning models. The second stream, which comprises an inception block, is utilized to capture multiscale features and the most complex features from the data through receptive parallel convolutional routes by using receptive fields. Two input tensors, which represent the residual and inception block routes, respectively, are used at the beginning of the algorithm. Each stream’s ensuing dilated convolutional and pooling layers are customized for the particular block type. For example, the first stream includes inception block with several parallel dilated convolutions for feature diversification, whereas the second stream includes residual block with unique shortcut connections and dilated convolutional processes. Incorporating dilated (or atrous) convolutional layers into the residual block and inception block improves the receptive field and collects contextual information. A typical residual block comprises a shortcut connection, also known as a skip connection, and two convolutional layers. Dilated convolutions can substitute or enhance the conventional convolutional layers in the residual block. The inclusion of dilated convolutions inside the residual block allows the network to effectively capture a broader context or global information while keeping the number of parameters relatively the same. By augmenting the dilation rate inside the designated residual block, the model concentrates on gathering information over a wider spectrum. Hence, it enhances its effectiveness for tasks necessitating extensive interdependencies. Let’s suppose X represents the input, F X identifies the shortcut connection, C o n v X , W ,   d i l a t i o n   r a t e shows the dilated convolution layers, and Y represents the output. Therefore, Equation (1) can be written as:
Y = F X + C o n v X ,   W ,   d i l a t i o n   r a t e ,
On the other hand, the inception block is composed of many concurrent branches, each applying a distinct processing method to the input. Typically, these branches consist of convolutional layers that have variable kernel sizes. Incorporating dilated convolutions into one of the branches of the inception block allows for a broader context to be considered during the process of feature extraction. Let’s suppose X is the input, C o n c a t C o n v 1 × 1 X ,   C o n v 3 × 3 X , C o n v d i l a t e d X , d i l a t i o n   r a t e , represents the different convolutional branches and dilated convolutional layers within the inception block, and Y is the output. So, mathematically, it can be expressed in Equation (2) as:
Y = C o n c a t C o n v 1 × 1 X ,   C o n v 3 × 3 X , C o n v d i l a t e d X , d i l a t i o n   r a t e ,
The network’s capability to collect characteristics at several scales enables it to effectively identify patterns of different dimensions. Dilated convolutional layers have an impact on both residual and inception blocks by allowing the network to gather contextual information from a wider region without adding much computational burden. It is of utmost importance for tasks that need the comprehension of both local and global features, such as image recognition. The use of dilated convolutions in these blocks varies depending on the architecture and the task’s unique demands.
Following the separate processing in both streams, a concatenation layer is used to combine the results to incorporate the unique features that each pathway has acquired. A technique for adaptively emphasizing significant characteristics from the concatenated representations is then presented. The soft-attention mechanism is used to enhance the model architecture, which comprises a pooling layer, dense layer, and reshaping, and the features are further refined by multiplication between attention weights and concatenated features, which improves the concentration of the proposed model on more prominent features [48,49]. The architecture mainly focuses on the separation and fusion of feature-extraction algorithms through the use of inception and residual blocks in parallel streams. Combining these disparate representations allows for a more thorough exploration of feature space, which is further improved by a soft-attention mechanism to allow for adaptive feature refining. To accommodate complex and varied feature representations for plant species classification tasks, the suggested architecture aims to maximize the special abilities of both residual and inception blocks. The principal objective is to increase classification performance in terms of accuracy and resilience. Figure 1 shows the general architecture of the proposed framework.

3.2. Soft-Attention Mechanism

Soft-attention mechanisms are crucial for improving the capabilities of neural networks as they allow the networks to selectively concentrate on particular areas of the input image. Within the realm of image classification, these methods are very important for identifying pertinent characteristics and patterns within an image. Our suggested architecture utilizes the soft-attention mechanism, which is applied strategically after combining the characteristics from the dual-stream residual block and the inception block. The soft-attention mechanism operates by allocating attention weights to distinct regions or channels of the input feature maps. Although our dual-stream architecture has the ability to extract features from the images using parallel blocks, it is not effective in each situation like in low lighting, which can affect the model’s capability of correctly classifying the plant species in indoor scenarios. Therefore, we utilized a soft-attention module, which effectively extracts the most pertinent features from the input images. The attention mechanism we utilized in this task consists of the PReLU activation function, max pooling layer, soft-attention unit, concatenation, and dropout layer in the last. The activation function assesses the input from the preceding layer and transmits its weights instantly to the soft-attention unit and pooling layer. The feature tensor b is used as the input in the soft-attention unit of the deep-learning model.
f s a m = a b k = 1 K s o f t m a x   W k . b ,
where the input of the 3D convolutional layer is b   R h × w × d , the weights are represented by W k   R h × w × d × N , and the 3D weights are denoted by N . The convolution result is normalized via the SoftMax function to produce K attention maps where K = 16. The attention maps are merged to create a single attention map that functions as a weighting factor α , as seen in Figure 2. Later, the variable α is multiplied by t to precisely transform the values for essential features, and a is utilized for further scaling since particular images need different a values, with a considered as a training parameter. A residual branch is created by combining the precisely scaled features f s a m from Equation (3) with the original features b . While training our model, the value of y is set to 0.01, which allows the network to progressively adjust its attention needs and transmit the weights to maximize pooling layer 1. The pooling layer 2 acquired the core features directly from the PReLU activation function. Pooling layer 1 and pooling layer 2 feature maps are combined and then inputted into the PReLU activation functions, followed by a dropout layer. The soft-attention block assists the network in prioritizing the plant that is far from the camera or under low light conditions. The dual-stream architecture, with the included soft-attention technique, is well-suited for the classification of indoor plant species.

4. Experimental Results

This section offers a concise overview of the experimental setup, followed by a detailed explanation of the datasets utilized for the experimental analysis. Additionally, we provide an in-depth exploration of both quantitative and qualitative analysis in terms of comparing the performance of the proposed network with other state-of-the-art (SOTA) techniques. For the experimental results, we implemented the proposed network on intel corei9 CPU with Nvidia GeForce 3070 GPU (Nvidia Corporation, Santa Clara, CA, USA) embedded with 8 GB memory. Additionally, we used a well-known framework named Keras as the front end and TensorFlow was utilized as the back end. Further, we reimplemented various competitive techniques on both datasets and trained them for 100 epochs with a batch size of 32, as well as SGD optimizer with 0.0001 learning rate and weight decay while using a 0.3 dropout. We also used early-stopping mechanisms on the basis of validation loss and to prevent the model from overfitting. All these optimal values for regularization techniques and hyperparameter tuning are selected after performing different experiments with different ranges of values. We performed many experiments by changing the values of dropout rate, weight decay, learning rate, and early-stopping mechanism to select the optimal values at which the model provides its best performance and accuracy. For the model evaluation, we used accuracy, precision, recall, and F1 score, which are very common evaluation metrics in image classification tasks. The purpose of this experimental study is to obtain a deeper understanding and comparison of proposed model with other baseline models. The ablation study examines how different elements collectively influence the performance, reliability, and interpretability of the model in plant species classification.

4.1. Dataset Description and Pre-Processing

This section provides a comprehensive description of the two datasets utilized in our experimental analysis to determine the efficacy of the proposed network against various competitive techniques. Initially, we conducted extensive experiments over Flower-299 dataset, which contains 299 different species. A detailed description of those species is available in [50]. This dataset composes 115,944 different images, with an average width of 271 pixels and height of 242 pixels. The images are distributed into different classes, where the average number of samples per label is 387 images, as detailed in Table 2. This variation in the number of images per class reflects the diversity of different classes within the dataset. Such diversity is essential for robust training of deep-learning models, allowing the model to generalize well across all classes. The Flower-299 dataset is considered to be widely used in the plant domain. However, the dataset contains several limitations that need to be addressed. One notable limitation is the potential class imbalance, where certain flower species may have significantly fewer images compared to others. This imbalance can lead to biased model training and reduced performance, especially for underrepresented classes. Additionally, the dataset may lack diversity in terms of environmental conditions, such as variations in lighting, background, and image quality. Such uniformity may hinder the model’s ability to generalize well to real-world scenarios where environmental conditions can vary widely.
To address these limitations and ensure robust model training and performance, we developed our custom dataset, which comprises 48 distinct categories of plant species. The first version of the data was acquired from a plant firm using a “Canon 450D + EFS 18–55mm” camera (Canon Oita Factory, Oita, Japan), while the details about the dataset samples are mentioned in Table 2. Figure 3 shows the samples taken from our proposed plant dataset. The plants were cultivated inside, and the images were captured using two 45W flexible LED lamps. The images were captured from a distance of roughly 120 cm, using angles of 0 degrees, 45 degrees, and 60 degrees. To standardize the picture formats, we converted all the acquired photographs to .jpg format during the pre-processing step. We modified the dimensions of all the images to 3 where the dimension refers to the measurements of the length and width of an image. The measurement is often conducted in pixels, and we ensure that each image has a consistent depth of 24 bits. Following this step, we apply some more pre-processing steps involving different augmentation techniques to increase the number of samples because the deep-learning model needs to be trained on a huge variety of samples so that it classifies each plant species accurately. We used the Albumentations library to create 10 separate augmentation pipelines. Each pipeline is designed to add various variations and perturbations into input images hence enhancing the dataset for training deep-learning models. The augmentation techniques used in these pipelines encompass several transformations designed to improve the resilience and flexibility of the model. Operations such as random rotations, flips, variations in brightness contrast, and the use of Contrast-Limited Adaptive Histogram Equalization (CLAHE) enhance the capacity to adapt to different lighting situations. In addition, this dataset includes blur effects, gamma changes, and simulated weather conditions such as rain, fog, sun flare, and shadows to enhance its diversity and robustness.

4.2. Evaluation Metrics

The study utilized several evaluation indicators to determine the robustness of the proposed network and various SOTA techniques. These indicators include P r e c i s i o n , P r e c i s i o n , F 1 - S c o r e , and A c c u r a c y , as the detailed descriptions are mentioned in the following references [51]. These evaluation metrics are calculated through True Positive ( T P ), True Negative ( T N ), False Positive ( F P ), and False Negative ( F N ). The mathematical formulation of evaluation metrics is mentioned in the following equations:
A c c u r a c y = T P + T N T P + F P + T N + F N
P r e c i s i o n = T P T P + F P
P r e c i s i o n = T P T P + F N
F 1 - S c o r e = 2 × P r e c i s i o n × R e c a l l   /   P r e c i s i o n + R e c a l l

4.3. Performance Evaluation Based on the Proposed Network

This section provides a brief comparison of the proposed network against various competitive techniques in terms of quantitative and qualitative analysis over two challenging datasets, as discussed in the subsequent subsection.

4.3.1. Quantitative Comparison of the Proposed Network with Other Techniques

This subsection presents a detailed comparison of the proposed network and various competitive techniques over two benchmarks using distinct evaluation indicators such as accuracy, precision, recall, and F1 score, as tabulated in Table 3. In the experimental analysis, the proposed dual stream with soft-attention network attained superior performance over custom dataset, representing 90.89%, 91.11%, 90.84%, and 90.85% for accuracy, precision, recall, and F1 score respectively. In addition, DenseNet121 is considered a second-best model for our custom dataset, which obtained 89.81% accuracy, 91.1% precision, 90.47% recall, and 90.53% F1 score respectively. The internal architecture of our dual-stream network utilizes a blend of residual and inception blocks, including an attention mechanism to emphasize significant characteristics. This strategy enables the model to effectively capture complex patterns, especially when dealing with challenging input data, leading to enhanced accuracy in classification. Unlike several competitive techniques that struggle with precision and recall, our proposed network demonstrates a balanced performance between accuracy and the capacity to accurately identify pertinent instances of plant species.
To further examine the effectiveness and scalability of the proposed network, we provided a detailed assessment of the network with benchmark Flower299 dataset using different evaluation indicators for instance accuracy, precision, recall, and F1 score. As given in Table 3, the proposed network demonstrates outstanding performance for accuracy, precision, recall, and F1 score, showcasing 88.87%, 88.98%, 88.64% and 88.92%, respectively. Further, the DensNet169 achieved promising performance for accuracy, precision, recall, and F1 score, which is 87.43%, 87.3%, 86.94%, and 86.95%, respectively. In addition, the DenseNet169 is considered the second-best network among others over Flower299 dataset. In short, the detailed experimental analysis of the proposed network over custom dataset, and a publicly available dataset, justified the robustness of the proposed network while handling the intricacies of the plant classification problem, positions it as a possible paradigm for other image classification domains.

4.3.2. Qualitative Analysis in Terms of Visualized Result

We conducted an investigation into the qualitative analysis of a suggested model that combines inception and residual blocks, followed by a soft-attention mechanism. This model shows substantial improvements in plant species categorization. The model efficiently extracts complete features from plant images by using the capabilities of inception blocks for collecting multi-scale characteristics and residual blocks for fast deep network training. Incorporating dilated convolutional layers significantly improves the model’s capacity to collect contextual information across different scales. The soft-attention mechanism is essential for directing attention toward the most relevant regions of the image, hence enhancing the model’s capacity to differentiate between similar plant species. The use of this attention mechanism enables the model to dynamically emphasize important characteristics, resulting in enhanced precision and resilience in categorization. The qualitative findings are visualized in Figure 4, demonstrating that the suggested architecture achieves superior classification accuracy and improved generalization across various plant species. This highlights its usefulness and promise for practical applications in plant biology and agriculture.

4.3.3. Discussion

This section demonstrates the experimental findings and discussion of our proposed dual-stream architecture enhanced by a soft-attention mechanism for plant species classification. With the addition of a soft-attention mechanism, our model improves the precision and comprehensibility of species classification. The experiments are performed on two different datasets, which have a wide range of plant species. We have conducted thorough experimentation using evaluation metrics like accuracy, precision, recall, and F1-score to assess the model performance. Furthermore, the discussion analyzes the advantages and possible drawbacks of the model, which provides a comprehensive assessment of the model’s performance for future improvement in the field of plant species classification.
The incorporation of residual and inception blocks provides an optimal combination of feature-extraction capabilities. Residual block, by using skip connections and the capacity to learn residual mappings, enhances the seamless transmission of information throughout the network, addressing the issue of vanishing gradients and allowing the model to effectively capture complex characteristics. The inception block enhances the receptive field by using filters of different sizes simultaneously, enabling the model to efficiently collect information at several scales. Furthermore, the integration of a soft-attention mechanism after the fusion of features from residual and inception blocks provides an additional level of flexibility to the model. The soft-attention mechanism selectively provides weights to certain regions of the feature maps, prioritizing areas that are essential for precise classification. The model’s flexibility improves its capacity to focus on key facts while excluding unnecessary or duplicate information. Figure 4 demonstrates the classification performance of our proposed model on testing samples.
The effectiveness of the proposed model is credited to the integration of these architectural components, which are skilled at capturing complex and unique features from the input data. The model accomplishes a high Accuracy, Precision, Recall, and F1-score by merging the advantages of residual and inception blocks, as well as including a soft-attention mechanism. This guarantees accurate and effective classification and comprehensive coverage of all plant species attributes. The model’s ability to extract complex features and allocate attention in a nuanced manner allows it to identify subtle patterns in images, resulting in improved performance compared to other benchmark models.

4.4. Ablation Study

We performed a detailed ablation study to select the most optimal model for plant species classification. Initially, we use the simple CNN model by utilizing the residual and inception blocks for classification. The CNN architecture is well-suited for plant species classification tasks. The process begins with a convolutional layer that has 64 filters and a 7 × 7 kernel size by providing the input structure of 224 × 224 × 3. The features are then downsampled using batch normalization and max pooling. Next, the network is made up of inception and residual blocks, which are both renowned for their ability to capture hierarchical characteristics. Shortcut connections are included to reduce the occurrence of vanishing gradient issues and promote a smoother gradient flow during training. To extract various features, the inception block uses max pooling and parallel convolutional procedures with varied kernel sizes. Average pooling is used after each block to flatten the representation of features for dense layers and minimize spatial dimensions. Densely connected nodes make up the fully connected layers, which lower dimensionality while retaining higher-level abstractions. To reduce overfitting, dropout regularization is used. Softmax activation is used in the last layer to provide probabilities for classes, where it varies from dataset to dataset, making it appropriate for multi-class plant species classification problems. To accomplish efficient feature extraction and classification, this kind of model architecture combines conventional convolutional layers with novel residual and inception blocks to obtain and train hierarchical representations from images.
Second architecture designed for plant species classification is implemented where network begins with an input layer that has dimensions of 224 × 224 × 3, which represent the width, height, and three RGB channels of the image. Next, it employs dilated convolutional layers with a 7 × 7 kernel and a stride of 2, followed by batch normalization and max-pool operations to effectively extract low-level features and minimize the data spatially. Residual block is incorporated into the model where each consists of two 3 × 3 convolutional layers that are equipped with skip connections and batch normalization. Then, inception block with their parallel dilated convolutional branches (1 × 1, 3 × 3, and 5 × 5 convolutions) and max-pooling functions allow for the extraction of features at different scales, which improves the network’s comprehension of complex images. The inception block is followed by an attention mechanism. To compute attention weights, this approach first uses dense layers, then global average pooling. To improve the network’s attention on critical image regions for classification, these weights are rearranged and multiplied elementwise with the feature maps. The goal is to amplify informative regions and decrease less relevant parts. The approach progresses from condensing feature maps by average pooling to flattening and completely correlated layers, which accomplish high-level feature abstraction. In the end, dropout regularization and Softmax layers are used for multi-class categorization over several classes, which completes the network’s architecture.
The third architecture consists of two separate paths that handle the input data through different phases of pooling and specific convolutional operations to enable different feature-extraction techniques. The initial step is to define the input tensors that represent the two parallel streams, residual block, and inception block. Each stream travels through a series of pooling and dilated convolutional layers that are specific to the block type it is assigned. The first stream, in particular, integrates residual block, which are distinguished by their capacity to maintain gradient flow and enable feature learning via shortcut connections. Parallelly, the second stream utilizes the inception block to obtain complicated features by using various receptive fields via parallel convolutional routes. The concatenate layer concatenates the features from both streams, whereas the concatenated feature representations from both streams enable the combination of complementary and varied features. This layer is followed by dense layers and output layers for the classification of plant species. The advantage of this architecture is that it utilizes the inception and residual block for enhancing the feature representation from two different streams where the gradient degradation is tackled by the residual block and feature hierarchies are captured through various receptive fields that is performed by the inception block. This approach attempts to learn more complimentary feature representation by combining the features from both parallel streams. This approach has the potential to improve the learning strategies in various computer vision applications and classification tasks by collecting a variety of features from different streams.

4.5. Drawbacks Associated with the Proposed Network

While the proposed network exhibited an optimal performance using two challenging datasets, as evident in the detailed experimental results, the model is associated with various drawbacks, as structured in the following points:
  • The proposed network comprises complex architecture, employing multiple streams followed by an attention mechanism. This complexity enhances its ability to capture complex patterns within input data, leading to improved overall performance. However, it can be challenging to manage and optimize such complex architecture, particularly during training.
  • Another limitation associated with the proposed model is high computational complexity, which can be challenging when deploying the model over edge devices for real-time decision-making. This limitation needs to be carefully considered when planning for deployment in resource-constrained environments.
  • The complexity of the network may lead to the risk of overfitting, particularly when training the network on limited or unbalanced datasets. Mitigating this risk may require additional regularization techniques or a larger and more diverse training dataset.

5. Conclusions and Future Directions

Our proposed dual-stream framework, followed by a soft-attention mechanism, demonstrates a strong and innovative system for classifying plant species where we utilized the residual block in the first stream and the inception block in the second stream. The utilization of dilated convolutional layers in both streams improves the model’s receptive field, making it capable of gathering features from a broader region in the images. The complex feature-extraction capacity is essential for identifying complex patterns and characteristics seen in plant species images. The combination of residual and inception blocks exploits the advantages of both architectures, enhancing the variety of features and producing detailed representations. Moreover, the residual block facilitates the mode of handling the vanishing gradient problem, while the inception block enhances the receptive field for extracting more prominent features from the data. The use of dilated convolutions also enhances the model’s capacity to extract characteristics at various sizes without substantially increasing computational complexity. Our proposed architecture has a soft-attention mechanism after the concatenation of feature maps from both streams, which is a distinctive characteristic of this architecture. The use of this soft attention enables the model to flexibly concentrate on more appropriate regions, hence enhancing its capability to extract more features. The attention process is essential for identifying small differences among different plant species, which contributes to the high accuracy in classification. Our suggested architecture has shown exceptional performance in plant species classification tasks via rigorous testing and assessment. Our model outperforms other baseline models in plant species classification for botanical applications because of the meticulous incorporation of many elements like the utilization of dilated convolutions in residual and inception blocks and the soft-attention mechanism. Our proposed model for plant science demonstrates the effectiveness of mixing distinct architectural aspects to enhance accuracy and generalization in recognizing plant species.
When considering the future trajectory of our suggested design, various intriguing paths might be explored and improved upon. Initially, conducting more research on enhancing hyperparameters and refining the model architecture might result in even more remarkable improvements in performance. Investigating the interpretability and explainability features of the model’s predictions might improve the model’s usefulness in real-world situations. Moreover, the expansion of the dual-stream design to include multi-modal inputs, such as extra spectral or temporal data, shows potential for thorough plant species classification. By integrating a variety of data sources, a comprehensive understanding of plant ecosystems may be achieved, allowing the model to accurately capture subtle patterns and changes. Gaining insight into the primary factors that influence a certain categorization choice may enhance the confidence and acceptance of the model in agricultural and ecological research. As technology progresses, the suggested framework in edge computing settings, such as agricultural drones or field-based sensors, has the potential to revolutionize real-time plant species identification. Creating lighter versions of the model and incorporating it onto edge devices might provide on-site, in-field applications with minimum delay.

Author Contributions

Conceptualization, I.U.K.; Methodology, I.U.K.; Validation, I.U.K. and J.W.L.; Formal analysis, J.W.L.; Investigation, J.W.L.; Resources, J.W.L.; Data curation, I.U.K., H.A.K. and J.W.L.; Writing—original draft, I.U.K.; Writing—review & editing, H.A.K.; Visualization, I.U.K.; Supervision, J.W.L.; Project administration, J.W.L.; Funding acquisition, J.W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-RS-2022-00156354) supervised by the IITP (Institute for Information and Communications Technology Planning & Evaluation). It was also supported by the Institute of Information and Communications Technology Planning and Evaluation (IITP) under the metaverse support program to nurture the best talents (IITP-2023-RS-2023-00254529) grant funded by the Korean government (MSIT).

Data Availability Statement

Data are contained within the article.

Acknowledgments

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-RS-2022-00156354) supervised by the IITP (Institute for Information and Communications Technology Planning & Evaluation). It was also supported by the Institute of Information and Communications Technology Planning and Evaluation (IITP) under the metaverse support program to nurture the best talents (IITP-2023-RS-2023-00254529) grant funded by the Korean government (MSIT).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Broich, M.; Huete, A.; Paget, M.; Ma, X.; Tulbure, M.; Coupe, N.R.; Evans, B.; Beringer, J.; Devadas, R.; Davies, K. A spatially explicit land surface phenology data product for science, monitoring and natural resources management applications. Environ. Model. Softw. 2015, 64, 191–204. [Google Scholar] [CrossRef]
  2. Cleland, E.E.; Chuine, I.; Menzel, A.; Mooney, H.A.; Schwartz, M.D. Shifting plant phenology in response to global change. Trends Ecol. Evol. 2007, 22, 357–365. [Google Scholar] [CrossRef]
  3. Richardson, A.D.; Keenan, T.F.; Migliavacca, M.; Ryu, Y.; Sonnentag, O.; Toomey, M. Climate change, phenology, and phenological control of vegetation feedbacks to the climate system. Agric. For. Meteorol. 2013, 169, 156–173. [Google Scholar] [CrossRef]
  4. Rehman, A.; Naz, S.; Razzak, M.I.; Akram, F.; Imran, M. A deep learning-based framework for automatic brain tumors classification using transfer learning. Circuits Syst. Signal Process. 2020, 39, 757–775. [Google Scholar] [CrossRef]
  5. Kaur, S.; Pandey, S.; Goel, S. Plants disease identification and classification through leaf images: A survey. Arch. Comput. Methods Eng. 2019, 26, 507–530. [Google Scholar] [CrossRef]
  6. Wu, S.G.; Bao, F.S.; Xu, E.Y.; Wang, Y.-X.; Chang, Y.-F.; Xiang, Q.-L. A leaf recognition algorithm for plant classification using probabilistic neural network. In Proceedings of the 2007 IEEE International Symposium on Signal Processing and Information Technology, Giza, Egypt, 15–18 December 2007; pp. 11–16. [Google Scholar]
  7. Tiwari, S. A comparative study of deep learning models with handcraft features and non-handcraft features for automatic plant species identification. Int. J. Agric. Environ. Inf. Syst. (IJAEIS) 2020, 11, 44–57. [Google Scholar] [CrossRef]
  8. Yang, K.; Zhong, W.; Li, F. Leaf segmentation and classification with a complicated background using deep learning. Agronomy 2020, 10, 1721. [Google Scholar] [CrossRef]
  9. Priya, C.A.; Balasaravanan, T.; Thanamani, A.S. An efficient leaf recognition algorithm for plant classification using support vector machine. In Proceedings of the International Conference on Pattern Recognition, Informatics and Medical Engineering (PRIME-2012), Salem, India, 21–23 March 2012; pp. 428–432. [Google Scholar]
  10. Du, J.-X.; Zhai, C.-M.; Wang, Q.-P. Recognition of plant leaf image based on fractal dimension features. Neurocomputing 2013, 116, 150–156. [Google Scholar] [CrossRef]
  11. Bodhwani, V.; Acharjya, D.; Bodhwani, U. Deep residual networks for plant identification. Procedia Comput. Sci. 2019, 152, 186–194. [Google Scholar] [CrossRef]
  12. Sobha, P.M.; Thomas, P.A. Deep learning for plant species classification survey. In Proceedings of the 2019 International Conference on Advances in Computing, Communication and Control (ICAC3), Mumbai, India, 20–21 December 2019; pp. 1–6. [Google Scholar]
  13. Weiss, U.; Biber, P.; Laible, S.; Bohlmann, K.; Zell, A. Plant species classification using a 3D LIDAR sensor and machine learning. In Proceedings of the 2010 Ninth International Conference on Machine Learning and Applications, Washington, DC, USA, 12–14 December 2010; pp. 339–345. [Google Scholar]
  14. Hussein, B.R.; Malik, O.A.; Ong, W.-H.; Slik, J.W.F. Automated classification of tropical plant species data based on machine learning techniques and leaf trait measurements. In Proceedings of the Computational Science and Technology: 6th ICCST 2019, Kota Kinabalu, Malaysia, 29–30 August 2019; Springer: Singapore, 2020; pp. 85–94. [Google Scholar]
  15. Wäldchen, J.; Mäder, P. Plant species identification using computer vision techniques: A systematic literature review. Arch. Comput. Methods Eng. 2018, 25, 507–543. [Google Scholar] [CrossRef]
  16. Du, J.-X.; Huang, D.-S.; Wang, X.-F.; Gu, X. Computer-aided plant species identification (CAPSI) based on leaf shape matching technique. Trans. Inst. Meas. Control 2006, 28, 275–285. [Google Scholar] [CrossRef]
  17. Wang, X.-F.; Huang, D.-S.; Du, J.-X.; Xu, H.; Heutte, L. Classification of plant leaf images with complicated background. Appl. Math. Comput. 2008, 205, 916–926. [Google Scholar] [CrossRef]
  18. Valliammal, N.; Geethalakshmi, S.N. An optimal feature subset selection for leaf analysis. Int. J. Comput. Inf. Eng. 2012, 6, 191–196. [Google Scholar]
  19. Wang, G.; Sun, Y.; Wang, J. Automatic image-based plant disease severity estimation using deep learning. Comput. Intell. Neurosci. 2017, 2017, 2917536. [Google Scholar] [CrossRef]
  20. Parez, S.; Dilshad, N.; Alghamdi, N.S.; Alanazi, T.M.; Lee, J.W. Visual intelligence in precision agriculture: Exploring plant disease detection via efficient vision transformers. Sensors 2023, 23, 6949. [Google Scholar] [CrossRef]
  21. Parez, S.; Dilshad, N.; Alanazi, T.M.; Lee, J.-W. Towards Sustainable Agricultural Systems: A Lightweight Deep Learning Model for Plant Disease Detection. Comput. Syst. Sci. Eng. 2023, 47, 515–536. [Google Scholar] [CrossRef]
  22. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  27. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  28. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  29. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using deep learning for image-based plant disease detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef]
  30. Dyrmann, M.; Karstoft, H.; Midtiby, H.S. Plant species classification using deep convolutional neural network. Biosyst. Eng. 2016, 151, 72–80. [Google Scholar] [CrossRef]
  31. Sharma, V.; Mishra, A.K.; Paliwal, S. Machine Learning Framework for Recognition and Classification of Plant Species: A Study using Digital Database. In Proceedings of the 24th International Conference on Distributed Computing and Networking, Kharagpur, India, 4–7 January 2023; pp. 407–413. [Google Scholar]
  32. Palanivel, N.; Jayapradha, J.; Boovaneswari, S.; Saravanan, C. Convolutional Neural Network Architecture for Accurate Plant Classification. In Technology Innovation Pillars for Industry 4.0; CRC Press: Boca Raton, FL, USA, 2024; pp. 50–64. [Google Scholar]
  33. Van Huy, L.; Tuong, N.H.; Van Hieu, N. Masked autoencoder-based self-supervised learning for forest plant classification. Cybern. Phys. 2024, 13, 32–41. [Google Scholar]
  34. Barré, P.; Stöver, B.C.; Müller, K.F.; Steinhage, V. LeafNet: A computer vision system for automatic plant species identification. Ecol. Inform. 2017, 40, 50–56. [Google Scholar] [CrossRef]
  35. Lee, S.H.; Chan, C.S.; Mayo, S.J.; Remagnino, P. How deep learning extracts and learns leaf features for plant classification. Pattern Recognit. 2017, 71, 1–13. [Google Scholar] [CrossRef]
  36. Haque, F.; Haque, S. Plant recognition system using leaf shape features and Minimum Euclidean distance. ICTACT J. Image Video Process. 2018, 9, 1919–1925. [Google Scholar] [CrossRef]
  37. Gao, Z.-Y.; Xie, H.-X.; Li, J.-F.; Liu, S.-L. Spatial-structure siamese network for plant identification. Int. J. Pattern Recognit. Artif. Intell. 2018, 32, 1850035. [Google Scholar] [CrossRef]
  38. Zhu, Y.; Sun, W.; Cao, X.; Wang, C.; Wu, D.; Yang, Y.; Ye, N. TA-CNN: Two-way attention models in deep convolutional neural network for plant recognition. Neurocomputing 2019, 365, 191–200. [Google Scholar] [CrossRef]
  39. Dileep, M.; Pournami, P. AyurLeaf: A deep learning approach for classification of medicinal plants. In Proceedings of the TENCON 2019—2019 IEEE Region 10 Conference (TENCON), Kerala, India, 17–20 October 2019; pp. 321–325. [Google Scholar]
  40. Duong-Trung, N.; Quach, L.-D.; Nguyen, C.-N. Learning deep transferability for several agricultural classification problems. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 58–67. [Google Scholar] [CrossRef]
  41. Liu, J.; Yang, S.; Cheng, Y.; Song, Z. Plant leaf classification based on deep learning. In Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; pp. 3165–3169. [Google Scholar]
  42. Silva, P.F.; Marcal, A.R.; da Silva, R.M.A. Evaluation of features for leaf discrimination. In Proceedings of the International Conference Image Analysis and Recognition, Póvoa de Varzim, Portugal, 26–28 June 2013; pp. 197–204. [Google Scholar]
  43. Villaruz, J.A. Deep convolutional neural network feature extraction for berry trees classification. J. Adv. Inf. Technol. 2021, 12, 226–233. [Google Scholar] [CrossRef]
  44. Thomkaew, J.; Intakosum, S. Plant Species Classification Using Leaf Edge. J. Image Graph. 2023, 11, 91–97. [Google Scholar] [CrossRef]
  45. Siddharth, T.; Kirar, B.S.; Agrawal, D.K. Plant Species Classification Using Transfer Learning by Pretrained Classifier VGG-19. arXiv 2022, arXiv:2209.03076. [Google Scholar]
  46. Devi, M.S.; Aruna, R.; Rajeswari, D.R.; Manogna, R.S. Conv2D Xception Adadelta Gradient Descent Learning Rate Deep learning Optimizer for Plant Species Classification. In Proceedings of the 2023 Third International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT), Bhilai, India, 5–6 January 2023; pp. 1–4. [Google Scholar]
  47. Pushpa, B.; Lakshmi, P. Deep Learning Model for Plant Species Classification Using Leaf Vein Features. In Proceedings of the 2022 International Conference on Augmented Intelligence and Sustainable Systems (ICAISS), Trichy, India, 24–26 November 2022; pp. 238–243. [Google Scholar]
  48. Yar, H.; Ullah, W.; Khan, Z.A.; Baik, S.W. An effective attention-based CNN model for fire detection in adverse weather conditions. ISPRS J. Photogramm. Remote Sens. 2023, 206, 335–346. [Google Scholar] [CrossRef]
  49. Yar, H.; Khan, Z.A.; Rida, I.; Ullah, W.; Kim, M.J.; Baik, S.W. An efficient deep learning architecture for effective fire detection in smart surveillance. Image Vis. Comput. 2024, 145, 104989. [Google Scholar] [CrossRef]
  50. Bogdan Cretu, A.I. Flowers-299. Available online: https://www.kaggle.com/datasets/bogdancretu/flower299 (accessed on 22 February 2024).
  51. Khan, I.U.; Afzal, S.; Lee, J.W. Human activity recognition via hybrid deep learning based model. Sensors 2022, 22, 323. [Google Scholar] [CrossRef]
Figure 1. A high-level generic diagram of the proposed study, comprising three distinct phases: Phase (1) contains different augmentation techniques to mitigate the data-scarcity problem. Phase (2) showcases a dual-stream network, employing two different networks, such as residual block and inception block, followed by a soft-attention module for effective feature learning. Finally, Phase (3) signifies the testing phase, where the proposed network is evaluated on testing data.
Figure 1. A high-level generic diagram of the proposed study, comprising three distinct phases: Phase (1) contains different augmentation techniques to mitigate the data-scarcity problem. Phase (2) showcases a dual-stream network, employing two different networks, such as residual block and inception block, followed by a soft-attention module for effective feature learning. Finally, Phase (3) signifies the testing phase, where the proposed network is evaluated on testing data.
Plants 13 02655 g001
Figure 2. The soft-attention function of converting feature map to attention maps and postprocessing procedure.
Figure 2. The soft-attention function of converting feature map to attention maps and postprocessing procedure.
Plants 13 02655 g002
Figure 3. Shows the different types of plant species samples from our proposed dataset.
Figure 3. Shows the different types of plant species samples from our proposed dataset.
Plants 13 02655 g003
Figure 4. Demonstrates the classification performance of our proposed model on testing samples, where (ac) shows samples from our custom dataset, while (df) shows samples from the Flower299 dataset.
Figure 4. Demonstrates the classification performance of our proposed model on testing samples, where (ac) shows samples from our custom dataset, while (df) shows samples from the Flower299 dataset.
Plants 13 02655 g004
Table 1. Representation of the summary of previous techniques used for plant classification.
Table 1. Representation of the summary of previous techniques used for plant classification.
RefDatasetsClassesModelsAccuracy
[7]Custom dataset30DNN91.17%
30CNN95.58%
[8]Custom dataset15VGG1691.50%
15VGG1992.40%
15Inception-ResNetV289.60%
[11]LeafSnap180ResNet93.09%
[29]PlantVillage38AlexNet99.27%
38GoogLeNet99.34%
[30]Six different datasets22CNN86.20%
[32]Vietnam Plant (VNP-200)200CNN-PC96.42%
[33]QuangNamForestPlant710Masked Autoencoder94%
[34]LeafSnap184LeafNet86.30%
60LeafNet95.80%
32LeafNet97.90%
[35]Malayakew44Deep CNN (D1) MLP97.70%
44Deep CNN (D1) SVM (linear)98.10%
[37]LifeCLEF 2015303SN84.20%
[39]AyurLeaf40AlexNet94.87%
40Ayurleaf CNN95.06%
[40]Custom dataset20MobileNet98.50%
[41]Flavia32Ten-layer CNN model87.92%
[43]Custom dataset3AlexNet97.80%
Table 2. Represents the summary of both datasets.
Table 2. Represents the summary of both datasets.
DatasetTotal ImagesTrainingTestingNo. of Classes
Flower299115,94492,75523,188299
Proposed Dataset93647492187248
Table 3. Comparative analysis of our proposed model against various competitive networks over custom dataset and Flower299.
Table 3. Comparative analysis of our proposed model against various competitive networks over custom dataset and Flower299.
No.ModelCustom DatasetFlower299 Dataset
Acc.Prec.Rec.F1Acc.Prec.Rec.F1
1DenseNet12189.8191.190.4790.5384.6984.4184.1584.16
2DensNet16985.2585.7384.7284.7687.4387.386.9486.95
3EffecientNetB078.1378.977.0378.1078.6778.178.8878.65
4InceptionResNet76.0678.8474.9575.0483.5584.4583.183.67
5InceptionV384.585.2883.7783.7784.5284.3684.0384.05
6ResNet5088.3588.988.1388.2384.7684.5384.2684.26
7MobileNetV285.8486.5785.2285.3785.1584.8984.6584.64
8VGG1688.990.6190.190.1984.3484.1683.8483.87
9VGG1988.388.9988.2988.3084.4884.2983.9784.01
10Inc-Res85.8486.5785.2285.3782.4882.7882.282.75
11Inc-Res SAM85.385.884.784.8183.1583.683.0283.45
12Dual Str. Inc-Res87.0987.3987.2987.1884.6585.283.984.78
13Proposed Model90.8991.1190.8490.8588.8788.9888.6788.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, I.U.; Khan, H.A.; Lee, J.W. Dual-Stream Architecture Enhanced by Soft-Attention Mechanism for Plant Species Classification. Plants 2024, 13, 2655. https://doi.org/10.3390/plants13182655

AMA Style

Khan IU, Khan HA, Lee JW. Dual-Stream Architecture Enhanced by Soft-Attention Mechanism for Plant Species Classification. Plants. 2024; 13(18):2655. https://doi.org/10.3390/plants13182655

Chicago/Turabian Style

Khan, Imran Ullah, Haseeb Ali Khan, and Jong Weon Lee. 2024. "Dual-Stream Architecture Enhanced by Soft-Attention Mechanism for Plant Species Classification" Plants 13, no. 18: 2655. https://doi.org/10.3390/plants13182655

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop