Next Article in Journal
Complete Chloroplast Genome of Bamboo Species Pleioblastus ovatoauritus and Comparative Analysis of Pleioblastus from China and Japan
Next Article in Special Issue
Learning from Forest Trees: Improving Urban Tree Biomass Functions
Previous Article in Journal
How Climate Change Impacts the Distribution of Lithocarpus hancei (Fagaceae), a Dominant Tree in East Asian Montane Cloud Forests
Previous Article in Special Issue
Allometric Growth of Common Urban Tree Species in Qingdao City of Eastern China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Species Identification in Urban Environments Using TensorFlow Lite and a Transfer Learning Approach

by
Diego Pacheco-Prado
1,2,*,
Esteban Bravo-López
2,3 and
Luis Ángel Ruiz
1
1
Geo-Environmental Cartography and Remote Sensing Group (CGAT), Universitat Politècnica de València, 46022 Valencia, Spain
2
Instituto de Estudios de Régimen Seccional del Ecuador (IERSE), Vicerrectorado de Investigaciones, Universidad del Azuay, Cuenca 010204, Ecuador
3
Centre for Advanced Studies in Earth Sciences, Energy and Environment, Department of Cartographic, Geodetic and Photogrammetric Engineering, University of Jaen, 23071 Jaen, Spain
*
Author to whom correspondence should be addressed.
Forests 2023, 14(5), 1050; https://doi.org/10.3390/f14051050
Submission received: 31 March 2023 / Revised: 24 April 2023 / Accepted: 15 May 2023 / Published: 19 May 2023
(This article belongs to the Special Issue Urban Forestry Measurements)

Abstract

:
Building and updating tree inventories is a challenging task for city administrators, requiring significant costs and the expertise of tree identification specialists. In Ecuador, only the Trees Inventory of Cuenca (TIC) contains this information, geolocated and integrated with the taxonomy, origin, leaf, and crown structure, phenological problems, and tree images taken with smartphones of each tree. From this dataset, we selected the fourteen classes with the most information and used the images to train a model, using a Transfer Learning approach, that could be deployed on mobile devices. Our results showed that the model based on ResNet V2 101 performed best, achieving an accuracy of 0.83 and kappa of 0.81 using the TensorFlow Lite interpreter, performing better results using the original model, with an accuracy and kappa of 0.912 and 0.905, respectively. The classes with the best performance were Ramo de novia, Sauce, and Cepillo blanco, which had the highest values of Precision, Recall, and F1-Score. The classes Eucalipto, Capuli, and Urapan were the most difficult to classify. Our study provides a model that can be deployed on Android smartphones, being the beginning of future implementations.

1. Introduction

Urban trees benefit citizens, improving their quality of life [1,2,3]; therefore, their conservation is essential for city administrators [1]. Urban Tree Inventories (UTIs) provide information about species and location and photography of each tree [2,3], which allow sustainable forest management and monitoring of urban trees and gardens [4]. Nevertheless, the costs of carrying out these inventories are usually high, impeding their creation [5] or updating [6], and requiring specialists to identify tree species [7]. Participatory projects such as Tree Inventory on Portland’s streets [8] or New York City [9], or private projects such as Cuenca-Ecuador [10] with smartphones [3,10,11] have allowed the collection of these datasets.
In recent years, identifying plant species [12,13,14,15,16], flowers [17], medicinal plants [18], and trees [19] has been achieved using photographs. Examples of this are some mobile applications available in the android market, such as LeafSnap, PictureThis, PlantNet, Blossom, Seek, and others [7]. Plant identification approaches are usually based on analyzing a single leaf followed by a flower [7]. For these image analysis systems that apply Artificial Vision and Deep Learning methods, it is possible to customize models of local flora [20], to assess the health status of trees and plants [16,21] and aid in leaf recognition and classification [22,23] using artificial neural networks.
Several studies have analyzed the application of Deep Learning for forest analysis from different approaches. In [24], Deep Learning algorithms were analyzed for forest resource inventory and tree species identification, implementing neural networks of different architectures. In [25], the use of Machine Learning (ML) algorithms (Random Forest, Support Vector Machines, and ANN Multilayer Perceptron) was analyzed for tree species classification in urban environments through the use and processing of LiDAR data. Although the focus of these studies is different from the one proposed in the present research, it is important to show that the use of quantitative methods based on ML algorithms contributes to the optimization of management, planning, and maintenance of tree inventories in urban areas with complex structures. On the other hand, there are also studies in which Deep Learning was applied explicitly for forest species identification. In [26], plant species mostly coming from western Europe and North America were identified based on images from the online collaborative Encyclopedia of Life https://eol.org/ (accessed on 20 February 2023), using Deep Convolutional Neural Network (DCNN) architectures, such as GoogleNet, ResNet, and ResNeXT. In [19], tree species were also identified using Convolutional Neural Networks (CNNs), but based on images of the bark. The images used were obtained under different conditions and cameras to ensure a diversification of the data. The accuracy obtained was higher than 93%, demonstrating the effectiveness of the application of this methodology. In [27], they used Deep Learning, implementing the UNET architecture (a variant of CNN), to classify tree species using forest images, considering additional important attributes for the stands and parcels from which the images were obtained, such as elevation, aspect, slope, and canopy density, and obtained accuracy results above 80% based on different validation metrics such as accuracy, Precision, and Recall. In [28], a more specific analysis was performed, detecting and identifying plant diseases in real-time using a CNN-based architecture on an integrated platform by analyzing the leaves. The accuracy of the results obtained was higher than 96%. All these investigations demonstrate that methodologies focused on Machine Learning and Deep Learning, such as CNNs, can be used to identify forest species from different perspectives and datasets with high accuracy.
CNN is nowadays the most used Neural Network for image classification. It consists of a stack of layers encompassing an input image, performing a mathematical operation, and predicting the class or label probabilities in the output [29]. This can be implemented in the TensorFlow framework [30], the most stable and complete library for the Python language [31].
Image classification models using CNNs can have millions of parameters, which means that training them from the beginning requires a large amount of labeled input data [19,32,33] and high computing power [34]. Another problem is that a limited number of samples can generate overfitting problems in CNN models, which consist of the poor generalization capacity of the models to new information [32]. This can be reduced through two techniques: Transfer Learning (TL) and Data Augmentation (DA). TL is a process that allows the creation of new learning models by adjusting once-trained neural networks [34,35]. DA generates synthetic samples from existing data using transformations, such as image rotation, color degradation, and others, although it can cause overfitting in minority classes [36]. TensorFlow Hub offers a pre-trained and ready-to-use model for TL [37] that simplifies training models using a custom dataset [38]. Additionally, the TensorFlow Lite Model Maker function allows for these models to be deployed for mobile devices.
The trained models can run inference on the client’s devices when using TFLite files for our Java-developed Android application. AgroAId is an example; it was built to identify combinations from a plant leaf image input, recognizing 39 species and disease classes on devices with limited computing and memory resources [39]. Finally, the relevant pixels for the decision to classify the image by CNN can be visualized through the GRAD-CAM (Gradient-weighted Class Activation Mapping) method, which highlights the areas of the image in the form of heat maps [32,40,41].
If we consider that most of the mobile applications available require an Internet connection to perform the identification of plants or forest species [7], that means that the inference is performed in a medium external to the mobile device and it limits its use in disconnected environments; thus our main contribution is to evaluate a methodology that trains a model that can be used offline and is based on Transfer Learning methodology due to the limited number of samples available compared to other free datasets. On the other hand, currently available mobile applications for plant identification focus on leaves or fruits because these elements are the most easily recognizable and distinctive parts of a plant. In the case of trees, the reviewed works evaluate photographs of the trunk or images with environmental noise. Nevertheless, our work analyzed trees based on their general appearance in order to identify those whose characteristics can be easily evaluated by a CNN, thus allowing for their identification.
This research applies and evaluates a Transfer Learning for automated identification of the most common tree species in the urban area of Cuenca (Ecuador), using computer vision techniques based on smartphone images of the local flora from the Tree Inventory of Cuenca (TIC), and can be deployed using Android smartphones.

2. Materials and Methods

2.1. Study Area

Cuenca is located in Ecuador’s central/southern area (Figure 1) at a mean elevation of 2500 m above sea level (m.a.s.l.). According to information from the Municipal Cleaning Company of Cuenca (EMAC-EP), in 2017 there were around 208 parks (public green areas) within the urban perimeter of the city [4], with the largest and the most important green areas in El Paraíso, Miraflores, and La Madre.
In Cuenca (Ecuador), the University of Azuay, through the Institute for Sectional Regime Studies of Ecuador—IERSE https://ierse.uazuay.edu.ec (accessed on 5 February 2023), —and the Azuay Herbarium https://herbario.uazuay.edu.ec/ (accessed on 5 February 2023), built the Tree Inventory of Cuenca and published the results on the official website https://gis.uazuay.edu.ec/iforestal/ (accessed on 1 March 2023), [10]. Until June 2022, it contained 12,688 trees with information about taxonomy, origin, crown and trunk structure, phenology, forest management, potential problems, and two photographs of each tree. Of these trees, 52.1% corresponded to introduced species, 47.7% are native, and less than 1% are endemic species. The most frequent species are Fresno o Cholan, Sauce, Urapan, Jaracaranda, Capuli, and Cepillo Blanco.

2.2. Methodology

The methodology was implemented in four phases (Figure 2): (1) assessment of existing classes in the TIC and selection of those with the largest number of images, with trees over two meters height. (2) Image selection by activation maps and balancing of the number of images per class using an image rotation operation to reach the minimum image number. (3) Training the models with the selected images, evaluation, and producing classification reports. (4) Deployment of the top-ranked TFLite model on Android devices.

2.2.1. Class Selection

Only those trees or shrubs over 2 m in height were selected from the TIC database. The Cucarda, Guaylo, and Tilo_or_sauco_blanco classes can be considered as shrubs [42]. A total of fourteen species were finally chosen (Table 1) based on the availability of at least 160 samples, with the highest number of inventoried trees of native (N) or introduced (I) origin.

2.2.2. Image Selection and Preparation

From the existing species in the TIC, the fourteen with the greatest number of individuals were selected. Of these, Cipres had the least number of images, so the selection process was performed based on the number of images of this class (426), 20% of which (86 images) were reserved for validation. From this set of images, a selection process was performed through activation maps (Figure 3) in order to use only the images where the relevant pixels (red color in Figure 3) belonged to the actual trees and not to the context (buildings, cars, and others). For each selected image, a corresponding activation map (new image) was generated using the EfficientNetB0 base model from TensorFlow Hub and the Grad-CAM library in Python version 3.9.12. Each generated activation map was visually inspected to determine which images to eliminate based on whether EfficientNetB0 considered the context pixels more relevant than the tree pixels.
Therefore, the number of images for training was reduced, and it was necessary to complete them through synthetic samples obtained by rotation at 90, 180, and 270 degrees (Table 2). This is Data augmentation (DA) which is a technique used to increase the size of the training data set by creating new images from existing ones [30]. Finally, the image size was homogenized through a Python script that performed framing from the center and resized the sample images into 224 × 224 pixel images.
The final photographs of the selected classes were taken between 2017 and 2019 between 07:00 and 18:00, with a higher frequency between 9:00 and 12:00. Approximately 98% of the photographs were acquired with Huawei and Samsung branded smartphones, and the remaining 2% with Apple and Sony phones.

2.2.3. Training and Testing Models

Once the training and testing images were defined, the process of model training was performed by TL, for which two base architectures with a standard input size of 224 × 224 pixels and 3 bands were selected: RestNet V2 with 101 layers, trained on ImageNet (ILSVRC-2012-CLS) [49], and EfficientNet-Lite, with several versions with different parameters and optimized for edge devices [50]. The two models were implemented under the TensorFlow Lite model maker library with 20 epochs, which represents the number of times that all training images are fed through the neural network [51]. In the training, we defined the URL of the base model used from TensorFlow Hub and the address of the folder with the validation images. Finally, we fine-tuned the whole model instead of just training the head layer [52].
Two approaches were used to compare the performance of the models with the evaluation dataset: the confusion matrix and the classification reports from the Scikit-learn library (Python). First, the accuracy and Cohen’s Kappa metrics allowed us to assess the global reliability of the model. In the second tool, the Precision, Recall, and F1-score (harmonic mean between Precision and Recall) [53] were used to evaluate the performance of each class individually. The equations of these metrics are shown in Equations (1)–(5).
A c c u r a c y = T P + T N T P + T N + F P + F N
K a p p a = ( p o p e ) ( 1 p e )
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 - s c o r e = ( 2 P r e c i s i o n R e c a l l ) R e c a l l + P r e c i s i o n
where TP is true positive, TN is true negative, FP is false positive, and FN is false negative in the confusion matrix [53]. For Cohen’s Kappa, p o is the empirical probability of agreement on the label assigned to any sample (the observed agreement ratio) and p e is the expected agreement when both annotators assign the labels randomly [54].

2.2.4. Deployment of the Model in Android Application

The Tensorflow Lite model maker library was chosen for its ease of use in generating models for mobile devices and its integration with the templates proposed in the TensorFlow website [55]. This library saves the trained model in TFLite (TensorFlow lite) format. This format is lightweight, efficient, optimized for mobile devices, and can be used to package trained TensorFlow models for deployment [56].
The mobile devices require an interpreter to execute these models, and uses a combination of hardware acceleration and software optimization to perform the inference based on input data [39,57].

3. Results

3.1. Learning Curves of the Models

The performance of the proposed models, EfficientNet-Lite and ResNet V2 101, for forest species recognition was tested with Transfer Learning. Figure 4 shows the accuracy and loss learning curves of the EfficientNet-Lite model during training and validation. The training accuracy gradually increased (Figure 4a) and the training and validation loss decreased slowly across all epochs (Figure 4b); in both cases, the accuracy showed fluctuations between epochs 4 and 14 and later stabilized. Figure 5 shows the accuracy and loss of the ResNet-V2-101-based model from epoch three onwards, where the accuracy of the validation data reached a score of 0.9; after which the model did not learn anymore.
Analyzing the accuracy and Cohen’s kappa and the Accuracy (Acc) obtained from the validation dataset, the EfficientNet-Lite model performed better than the ResNet V2 101 model (Table 3). We evaluated the model and interpreter statistics in two ways. The first method, called Full Training, involved training the entire model, including layers and weights, which consumed a significant number of computational resources and time. The second method, called Top Layer Training, only retrained the top layers of the model, and utilized the previously learned weights and features for the other layers.
We assessed the predictions made by the TensorFlow Lite Interpreter, and we found that ResNet V2 101 was the best choice, even though its reliability decreased compared to the original model.

3.2. Class Accuracy with the Best Model

Attending to the F1-score of the model based on ResNet V2 101, there were four highly reliable classes: Ramo de novia, Cipres, Cepillo blanco, and Sauce (Figure 6). The loss of F1-score between the model and the interpreter was not homogeneous across all classes. Note that it was higher for the Cepillo and Urapan classes compared to the others.
Using the full report of the classification model (Table 4) and the interpreter (Table 5), a total of eight classes exceeded an F1-score of 0.8. The classes Eucalipto, Capuli, and Urapan were the most difficult to classify. The confusion matrices (model and interpreter) with the greatest confusion occurred between the Eucalipto and Capuli classes. In the case of the interpreter, in addition to the previous one, other cases also occurred, such as the confusion of Urapan with Eucalipto, and Cepillo with Cucarda.
Evaluating the classification probabilities assigned to each of the 86 images per class (Figure 7), the classes with the best F1-score ranking also presented a high probability of correctness. The least favorable probabilities were for Eucalipto, which tended to be confused with the Capuli class. For the Urapan class, the probability of success was lower.

3.3. Model Deployment in Android Devices

The mobile application installer and model in mobile format (TFLite) can be downloaded from the website https://gis.uazuay.edu.ec/ (accessed on 15 March 2023). When performing the real-time prediction, the tree characteristics vary due to wind and light effects; the four highest-ranking possibilities are presented in the screen. When similar values exist, the tree cannot be classified correctly at that moment, and a change of the camera position is needed. When there is a significant difference between the first and second predictions, then there is more certainty in the result. Additionally, the proximity of the camera sensor to the object influences the prediction, as in the example shown in Figure 8, where the probability of the prediction varied from 51.2% to 89.5%.

4. Discussion

In this work, two different types of results were obtained, from a model and interpreter (using smartphones). In both, the best accuracy and kappa metrics were obtained using ResNet V2 101 as the base model (Transfer Learning approach), with values of 0.912 and 0.905 in the model and 0.801 and 0.785 in the interpreter, respectively. This accuracy is similar to those obtained in works such as [39,53,58]. The TensorFlow Lite format, used in the interpreter, decreased the number of operations and affected the accuracy of the interpreter model.
The main difference between this work and others is that the images of TIC were taken entirely with smartphones, generally of the whole tree or shrub, and under different circumstances that add a high variability such as shadows, amount of light, distance to the object, and time of day, among others. For example, in [39], the images of leaves were taken using a Sony DSC—Rx100/13 and used for the identification of plant diseases. These images were at a resolution of 20.2 megapixels. On the other hand, the images in [58] focused on specific characteristics such as leaves and bark from different species.
In Transfer Learning [28,39], it was concluded that our dataset’s best-performing model for Transfer Learning was ResNet V2 101 that was fully retrained [59] (not only the classification layer on top). ResNet is an architecture that was previously used in similar applications such as [39] with 50 layers, while the one used in this work was 101 layers deep. From existing ResNet architectures, ResNet V2 101 had high reliability in other works with small datasets for the classification of plant leaf diseases [22] or in works using drone images, such as [59,60,61]. However, the ResNet V2 101 model has many more parameters than simpler models [62], which significantly increases the memory and processing requirements needed to train the model. In this work, controlling the number of images used for training was necessary due to computational constraints such as limited memory, CPU usage, and non-GPU availability, as well as processor capacity. In order to limit the number of images used, we chose to use only the image rotation data augmentation technique, used in other works such as [33,39].
Using the F1-score as a reference, from the fourteen classes evaluated in the model, ten exceeded a value of 0.9 and four in the interpreter. In the model, the classes with the highest value of F1-score were Guaylo, Cepillo blanco, Cipres, Sauce, and Ramo de Novia. For the interpreter, the best classes were Cipres, Cepillo blanco, Sauce, and Ramo de Novia.
The morphological characteristics of the Ramo de Novia class differentiated it from the other evaluated species, making this the most reliable class. The classes with the lowest reliability were Eucalipto and Capuli in the model, and Urapan and Eucalipto in the interpreter. As a test, the Capuli class was re-evaluated with another set of training and validation photographs to determine if this factor affected its reliability, obtaining similar results to the original classification. The morphological characteristics of this tree make it susceptible to confusion with other species. For this class, we propose to specifically evaluate the leaves [63], which are the organ most used in the identification of plants, or fruits for their identification. Other options to explore are the use of images of flowers [7] or bark [19].
These highest and lowest accuracy classes will be analyzed in future work using feature extraction such as color, texture, shape [64], which are commonly used for leaf feature extraction, to describe the characteristics that make them identifiable to a CNN. In addition, other data augmentation techniques may be included such as Laplacian sharpening, Gaussian blur augmentation, contrast enhancement, shifting, cropping, and zooming [65,66].
Even though the model’s reliability was greater than that of the interpreter, we believe that further work should be able to improve the interpreter model, for example, using TensorFlow Model Optimization Toolkit [67]. This would be desirable in those circumstances where tree identification must be carried out in real time and the mobile device does not have Internet connection.

5. Conclusions

This paper presented an evaluation of two base models, ResNet V2 and EfficientNet-Lite, in a TensorFlow Lite model maker library for the recognition of fourteen classes of trees using the images of the TIC dataset. The performance of model with ResNet V2 101 was superior to that obtained with EfficientNet-Lite and the species Cepillo blanco, Cipres, Sauce, Guabisay, and Ramo de novia presented the highest values of accuracy, Kappa, and F1-score in the classification. The final model had an accuracy of 0.912 and Kappa of 0.905, and with the TensorFlow Lite interpreter having an accuracy of 0.801 and Kappa of 0.785.
In the future, new tree species will be tested so they can be added to the application, generating a new classification model. It is expected that the mobile application will help non-expert users in the identification of tree species through smartphones, and it is considered a starting point for the creation of a support for tree’s inventories generation and maintenance in urban environments.

Author Contributions

Conceptualization, D.P.-P. and L.Á.R.; methodology, D.P.-P.; investigation, D.P.-P. and E.B.-L.; writing—original draft preparation, D.P.-P. and E.B.-L.; supervision, L.Á.R. All authors have read and agreed to the published version of the manuscript. Authorship is limited to those who have contributed substantially to the work reported.

Funding

This research was funded by the University of Azuay in the context of investigation project 2020-0125 denominated “Caracterización de unidades forestales a partir de datos espectrales, espaciales y de relieve a distintas escalas. Aplicación a los bosques andinos del cantón Cuenca (Ecuador), Fase 2”.

Data Availability Statement

These data are a private dataset and can be solicited from the Instituto de Estudios de Regimen Seccional del Ecuador (IERSE) at http://ierse.uazuay.edu.ec (accessed on 30 March 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Salam, A. Internet of Things for Sustainable Forestry. In Internet of Things; Springer: Cham, Switzerland, 2020; pp. 243–271. ISBN 9783030352905. [Google Scholar]
  2. Solomou, A.D.; Topalidou, E.T.; Germani, R.; Argiri, A.; Karetsos, G. Importance, Utilization and Health of Urban Forests: A Review. Not. Bot. Horti Agrobot. Cluj-Napoca 2019, 47, 10–16. [Google Scholar] [CrossRef]
  3. Pacheco, D.; Ávila, L. Inventario de Parques y Jardines de La Ciudad de Cuenca Con UAV y Smartphones. In Proceedings of the XVI Conferencia de Sistemas de Información Geográfica, Cuenca, Spain, 27–29 September 2017; pp. 173–179. [Google Scholar]
  4. Pacheco, D. Drones En Espacios Urbanos: Caso de Estudio En Parques, Jardines y Patrimonio. Estoa 2017, 11, 159–168. [Google Scholar] [CrossRef]
  5. Das, S.; Sun, Q.C.; Zhou, H. GeoAI to Implement an Individual Tree Inventory: Framework and Application of Heat Mitigation. Urban For. Urban Green. 2022, 74, 127634. [Google Scholar] [CrossRef]
  6. Branson, S.; Wegner, J.D.; Hall, D.; Lang, N.; Schindler, K.; Perona, P. From Google Maps to a Fine-Grained Catalog of Street Trees. ISPRS J. Photogramm. Remote Sens. 2018, 135, 13–30. [Google Scholar] [CrossRef]
  7. Wäldchen, J.; Mäder, P. Plant Species Identification Using Computer Vision Techniques: A Systematic Literature Review. Arch Comput. Method E 2018, 25, 507–543. [Google Scholar] [CrossRef] [PubMed]
  8. Tree Inventory Project|Portland. Available online: https://www.portland.gov/trees/get-involved/treeinventory (accessed on 11 November 2022).
  9. Volunteers Count Every Street Tree in New York City|US Forest Service. Available online: https://www.fs.usda.gov/features/volunteers-count-every-street-tree-new-york-city-0 (accessed on 11 November 2022).
  10. Ávila Pozo, L.A. Implementación de un Sistema de Inventario Forestal de Parques Urbanos en la Ciudad de Cuenca. Universidad-Verdad 2017, 73, 79–89. [Google Scholar] [CrossRef]
  11. Nielsen, A.B.; Östberg, J.; Delshammar, T. Review of Urban Tree Inventory Methods Used to Collect Data at Single-Tree Level. Arboric. Urban For. 2014, 40, 96–111. [Google Scholar] [CrossRef]
  12. Wäldchen, J.; Mäder, P. Machine Learning for Image Based Species Identification. Methods Ecol. Evol. 2018, 9, 2216–2225. [Google Scholar] [CrossRef]
  13. Goyal, N.; Kumar, N.; Gupta, K. Lower-Dimensional Intrinsic Structural Representation of Leaf Images and Plant Recognition. Signal, Image Video Process. 2022, 16, 203–210. [Google Scholar] [CrossRef]
  14. Azlah, M.A.F.; Chua, L.S.; Rahmad, F.R.; Abdullah, F.I.; Alwi, S.R.W. Review on Techniques for Plant Leaf Classification and Recognition. Computers 2019, 8, 77. [Google Scholar] [CrossRef]
  15. Kaya, A.; Keceli, A.S.; Catal, C.; Yalic, H.Y.; Temucin, H.; Tekinerdogan, B. Analysis of Transfer Learning for Deep Neural Network Based Plant Classification Models. Comput. Electron. Agric. 2019, 158, 20–29. [Google Scholar] [CrossRef]
  16. Rodríguez-Puerta, F.; Barrera, C.; García, B.; Pérez-Rodríguez, F.; García-Pedrero, A.M. Mapping Tree Canopy in Urban Environments Using Point Clouds from Airborne Laser Scanning and Street Level Imagery. Sensors 2022, 22, 3269. [Google Scholar] [CrossRef]
  17. Xia, X.; Xu, C.; Nan, B. Inception-v3 for Flower Classification. In Proceedings of the 2nd International Conference on Image, Vision and Computing ICIVC 2017, Chengdu, China, 2–4 June 2017; pp. 783–787. [Google Scholar] [CrossRef]
  18. Leena Rani, A.; Devika, G.; Vinutha, H.R.; Karegowda, A.G.; Vidya, S.; Bhat, S. Identification of Medicinal Leaves Using State of Art Deep Learning Techniques. In Proceedings of the 2022 IEEE International Conference on Distributed Computing and Electrical Circuits and Electronics ICDCECE 2022, Ballari, India, 23–24 April 2022; pp. 11–15. [Google Scholar] [CrossRef]
  19. Carpentier, M.; Giguere, P.; Gaudreault, J. Tree Species Identification from Bark Images Using Convolutional Neural Networks. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018; pp. 1075–1081. [Google Scholar] [CrossRef]
  20. Muñoz Villalobos, I.A.; Bolt, A. Diseño y Desarrollo de Aplicación Móvil Para La Clasificación de Flora Nativa Chilena Utilizando Redes Neuronales Convolucionales. AtoZ Novas Práticas Inf. Conhecimento 2022, 11, 1–13. [Google Scholar] [CrossRef]
  21. Adedoja, A.; Owolawi, P.A.; Mapayi, T. Deep Learning Based on NASNet for Plant Disease Recognition Using Leave Images. In Proceedings of the icABCD 2019—2nd International Conference on Advances in Big Data, Computing and Data Communication Systems, Winterton, South Africa, 5–6 August 2019. [Google Scholar] [CrossRef]
  22. Zhang, R.; Zhu, Y.; Ge, Z.; Mu, H.; Qi, D.; Ni, H. Transfer Learning for Leaf Small Dataset Using Improved ResNet50 Network with Mixed Activation Functions. Forests 2022, 13, 2072. [Google Scholar] [CrossRef]
  23. Vilasini, M.; Ramamoorthy, P. CNN Approaches for Classification of Indian Leaf Species Using Smartphones. Comput. Mater. Contin. 2020, 62, 1445–1472. [Google Scholar] [CrossRef]
  24. Wang, Y.; Zhang, W.; Gao, R.; Jin, Z.; Wang, X. Recent Advances in the Application of Deep Learning Methods to Forestry. Wood Sci. Technol. 2021, 55, 1171–1202. [Google Scholar] [CrossRef]
  25. Cetin, Z.; Yastikli, N. The Use of Machine Learning Algorithms in Urban Tree Species Classification. ISPRS Int. J. Geo-Inf. 2022, 11, 226. [Google Scholar] [CrossRef]
  26. Lasseck, M. Image-Based Plant Species Identification with Deep Convolutional Neural Networks. In Proceedings of the CEUR Workshop Proceedings, Bloomington, IN, USA, 20–21 January 2017; Volume 1866. [Google Scholar]
  27. Liu, J.; Wang, X.; Wang, T. Classification of Tree Species and Stock Volume Estimation in Ground Forest Images Using Deep Learning. Comput. Electron. Agric. 2019, 166, 105012. [Google Scholar] [CrossRef]
  28. Gajjar, R.; Gajjar, N.; Thakor, V.J.; Patel, N.P.; Ruparelia, S. Real-Time Detection and Identification of Plant Leaf Diseases Using Convolutional Neural Networks on an Embedded Platform. Vis. Comput. 2022, 38, 2923–2938. [Google Scholar] [CrossRef]
  29. Gogul, I.; Kumar, V.S. Flower Species Recognition System Using Convolution Neural Networks and Transfer Learning. In Proceedings of the 4th Fourth International Conference on Signal Processing, Communication and Networking ICSCN 2017, Chennai, India, 16–18 March 2017; pp. 1–6. [Google Scholar] [CrossRef]
  30. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in Vegetation Remote Sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  31. Pang, B.; Nijkamp, E.; Wu, Y.N. Deep Learning With TensorFlow: A Review. J. Educ. Behav. Stat. 2020, 45, 227–248. [Google Scholar] [CrossRef]
  32. Kim, T.K.; Kim, S.; Won, M.; Lim, J.H.; Yoon, S.; Jang, K.; Lee, K.H.; Park, Y.D.; Kim, H.S. Utilizing Machine Learning for Detecting Flowering in Mid—Range Digital Repeat Photography. Ecol. Modell. 2021, 440, 109419. [Google Scholar] [CrossRef]
  33. Figueroa-Mata, G.; Mata-Montero, E. Using a Convolutional Siamese Network for Image-Based Plant Species Identification with Small Datasets. Biomimetics 2020, 5, 8. [Google Scholar] [CrossRef]
  34. Paper, D. Simple Transfer Learning with TensorFlow Hub. In State-of-the-Art Deep Learning Models in TensorFlow; Apress: Berkeley, CA, USA, 2021; pp. 153–169. [Google Scholar]
  35. Creador de Modelos TensorFlow Lite. Available online: https://www.tensorflow.org/lite/models/modify/model_maker (accessed on 4 August 2022).
  36. Shorten, C.; Khoshgoftaar, T.M. A Survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  37. TensorFlow Hub. Available online: https://www.tensorflow.org/hub (accessed on 18 January 2023).
  38. TensorFlow Lite Model Maker. Available online: https://www.tensorflow.org/lite/models/modify/model_maker (accessed on 18 January 2023).
  39. Reda, M.; Suwwan, R.; Alkafri, S.; Rashed, Y.; Shanableh, T. AgroAId: A Mobile App System for Visual Classification of Plant Species and Diseases Using Deep Learning and TensorFlow Lite. Informatics 2022, 9, 55. [Google Scholar] [CrossRef]
  40. Kumar, A.; Singh, S.B.; Satapathy, S.C.; Rout, M. MOSQUITO-NET: A Deep Learning Based CADx System for Malaria Diagnosis along with Model Interpretation Using GradCam and Class Activation Maps. Expert Syst. 2022, 39, e12695. [Google Scholar] [CrossRef]
  41. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-Cam: Why Did You Say That? Visual Explanations from Deep Networks via Gradient-Based Localization. Rev. Hosp. Clínicas 2016, 17, 331–336. [Google Scholar]
  42. Minga, D.; Verdugo, A. Árboles y Arbustos de Los Ríos de Cuenca; Universidad del Azuay: Cuenca, Ecuador, 2016; ISBN 978-9978-325-42-1. [Google Scholar]
  43. Populus alba in Flora of China@Efloras. Available online: http://www.efloras.org/florataxon.aspx?flora_id=2&taxon_id=200005643 (accessed on 10 May 2022).
  44. Sánchez, J. Flora Ornamental Española. Available online: http://www.arbolesornamentales.es/ (accessed on 15 October 2022).
  45. Mahecha, G.; Ovalle, A.; Camelo, D.; Rozo, A.; Barrero, D. Vegetación Del Territorio CAR, 450 Especies de Sus Llanuras y Montañas; Corporación Autónoma Regional de Cundinamarca (Bogotá): Bogotá, Colombia, 2004; ISBN 958-8188-06-7. [Google Scholar]
  46. Guillot, D.; Van der Meer, P. El Género Yucca L En España; Jolube: Jaca, Spain, 2009; Volume 2, ISBN 978-84-937291-8-9. [Google Scholar]
  47. Tilo (Sambucus Canadensis) · INaturalist Ecuador. Available online: https://ecuador.inaturalist.org/taxa/84300-Sambucus-canadensis (accessed on 4 January 2023).
  48. Fresno (Fraxinus Uhdei) · INaturalist Ecuador. Available online: https://ecuador.inaturalist.org/taxa/134212-Fraxinus-uhdei (accessed on 4 January 2023).
  49. ResNet V2 101|TensorFlow Hub. Available online: https://tfhub.dev/google/imagenet/resnet_v2_101/feature_vector/5 (accessed on 6 February 2023).
  50. Ab Wahab, M.N.; Nazir, A.; Ren, A.T.Z.; Noor, M.H.M.; Akbar, M.F.; Mohamed, A.S.A. Efficientnet-Lite and Hybrid CNN-KNN Implementation for Facial Expression Recognition on Raspberry Pi. IEEE Access 2021, 9, 134065–134080. [Google Scholar] [CrossRef]
  51. Mezgec, S.; Seljak, B.K. Nutrinet: A Deep Learning Food and Drink Image Recognition System for Dietary Assessment. Nutrients 2017, 9, 657. [Google Scholar] [CrossRef]
  52. Object Detection with TensorFlow Lite Model Maker. Available online: https://www.tensorflow.org/lite/models/modify/model_maker/object_detection (accessed on 6 February 2023).
  53. Zheng, H.; Sherazi, S.W.A.; Son, S.H.; Lee, J.Y. A Deep Convolutional Neural Network-based Multi-class Image Classification for Automatic Wafer Map Failure Recognition in Semiconductor Manufacturing. Appl. Sci. 2021, 11, 9769. [Google Scholar] [CrossRef]
  54. Sklearn.Metrics.Cohen_kappa_score—Scikit-Learn 1.2.1 Documentation. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.metrics.cohen_kappa_score.html (accessed on 2 January 2023).
  55. Recognize Flowers with TensorFlow Lite on Android. Available online: https://codelabs.developers.google.com/codelabs/recognize-flowers-with-tensorflow-on-android#0 (accessed on 6 September 2022).
  56. Shah, V.; Sajnani, N. Multi-Class Image Classification Using CNN and Tflite. Int. J. Res. Eng. Sci. Manag. 2020, 3, 65–68. [Google Scholar] [CrossRef]
  57. TensorFlow Lite Inference. Available online: https://www.tensorflow.org/lite/guide/inference (accessed on 16 February 2023).
  58. Homan, D.; du Preez, J.A. Automated Feature-Specific Tree Species Identification from Natural Images Using Deep Semi-Supervised Learning. Ecol. Inform. 2021, 66, 101475. [Google Scholar] [CrossRef]
  59. Natesan, S.; Armenakis, C.; Vepakomma, U. Resnet-Based Tree Species Classification Using Uav Images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci.—ISPRS Arch. 2019, 42, 475–481. [Google Scholar] [CrossRef]
  60. Dixit, A.; Nain Chi, Y. Classification and Recognition of Urban Tree Defects in a Small Dataset Using Convolutional Neural Network, Resnet-50 Architecture, and Data Augmentation. J. For. 2021, 8, 61–70. [Google Scholar] [CrossRef]
  61. Zhang, C.; Xia, K.; Feng, H.; Yang, Y.; Du, X. Tree Species Classification Using Deep Learning and RGB Optical Images Obtained by an Unmanned Aerial Vehicle. J. For. Res. 2021, 32, 1879–1888. [Google Scholar] [CrossRef]
  62. Chen, L.; Li, S.; Bai, Q.; Yang, J.; Jiang, S.; Miao, Y. Review of Image Classification Algorithms Based on Convolutional Neural Networks. Remote Sens. 2021, 13, 4712. [Google Scholar] [CrossRef]
  63. Suwais, K.; Alheeti, K.; Al_Dosary, D. A Review on Classification Methods for Plants Leaves Recognition. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 92–100. [Google Scholar] [CrossRef]
  64. Shaji, A.P.; Hemalatha, S. Data Augmentation for Improving Rice Leaf Disease Classification on Residual Network Architecture. In Proceedings of the 2nd International Conference on Advances in Computing, Communication and Applied Informatics ACCAI 2022, Chennai, India, 28–29 January 2022. [Google Scholar] [CrossRef]
  65. Mikołajczyk, A.; Grochowski, M. Data Augmentation for Improving Deep Learning in Image Classification Problem. In Proceedings of the 2018 International Interdisciplinary PhD Workshop IIPhDW 2018, Swinoujscie, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar] [CrossRef]
  66. Barrientos, J.L. Leaf Recognition with Deep Learning and Keras Using GPU Computing; Engineering School, University Autonomous of Barcelona: Barcelona, Spain, 2018. [Google Scholar]
  67. TensorFlow Model Optimization|TensorFlow Lite. Available online: https://www.tensorflow.org/lite/performance/model_optimization (accessed on 22 February 2023).
Figure 1. Location of the city of Cuenca and its parks.
Figure 1. Location of the city of Cuenca and its parks.
Forests 14 01050 g001
Figure 2. Overlay workflow of the methodology followed to create image classification model and deployment on Android devices.
Figure 2. Overlay workflow of the methodology followed to create image classification model and deployment on Android devices.
Forests 14 01050 g002
Figure 3. Examples of selected and unselected images from their activation maps using Grad-CAM and Inception V3.
Figure 3. Examples of selected and unselected images from their activation maps using Grad-CAM and Inception V3.
Forests 14 01050 g003
Figure 4. The learning curves of the EfficientNet-Lite model using Transfer Learning: (a) accuracy learning curves; (b) loss learning curves.
Figure 4. The learning curves of the EfficientNet-Lite model using Transfer Learning: (a) accuracy learning curves; (b) loss learning curves.
Forests 14 01050 g004
Figure 5. Learning curves of the ResNet V2 101 model using Transfer Learning: (a) accuracy learning curves; (b) loss learning curves.
Figure 5. Learning curves of the ResNet V2 101 model using Transfer Learning: (a) accuracy learning curves; (b) loss learning curves.
Forests 14 01050 g005
Figure 6. F1-scores for all classes considered using ResNet V2 101 model base.
Figure 6. F1-scores for all classes considered using ResNet V2 101 model base.
Forests 14 01050 g006
Figure 7. Sum of probabilities by classes of the 86 testing images per class.
Figure 7. Sum of probabilities by classes of the 86 testing images per class.
Forests 14 01050 g007
Figure 8. Real-time example classification of the Ramo de Novia class. (a) At a distance of approximately one meter, and (b) at a distance of approximately five meters.
Figure 8. Real-time example classification of the Ramo de Novia class. (a) At a distance of approximately one meter, and (b) at a distance of approximately five meters.
Forests 14 01050 g008
Table 1. Tree species/class selected from the TIC database. (N) Native and (I) introduced species.
Table 1. Tree species/class selected from the TIC database. (N) Native and (I) introduced species.
Class NameDescriptionOrigin
Alamo blanco
Populus alba L.
Tree that reaches up to 15 m in height, with straight trunks and greenish white outer bark that cracks over the years [43].IForests 14 01050 i001
Capuli
Prunus serotina Ehrh.
Tree that reaches from 8 to 15 m in height and 30 to 50 cm in diameter at breast height (DBH). It has fissured outer bark, alternate branching, and a globose crown [42].NForests 14 01050 i002
Cepillo
Callistemon lanceolatus (Sm.) Sweet
Tree from 2 to 7 m in height with dark brown fissured outer bark. It has elongated leaves with a linear shape that are pubescent when young [42].IForests 14 01050 i003
Cepillo blanco
Melaleuca armillaris (Sol. ex Gaertn.) Sm.
Tree from 2 to 6 m in height with straight trunk and grayish brown fibrous outer bark. It has very thin leaves of linear shape with an acute apex [44].IForests 14 01050 i004
Cipres
Cupressus macrocarpa Hartw.
They can reach 20 m in height with an approximate diameter of about 60 cm. They have a straight trunk and thin bark with longitudinal fissures. Their leaves are very small, scale-like, and aligned in opposite pairs [42].IForests 14 01050 i005
Cucarda
Hibiscus rosa-sinensis L.
Shrub with branched stems from 2 to 5 m in height. It has simple leaves that are rounded at the base and elongated towards the apex. Its flowers can be presented in various colors, most commonly red or pink [42].IForests 14 01050 i006
Eucalipto
Eucaliptus globulus Labill.
They can reach more than 60 m in height. In some specimens, the outer bark is light brown with a skin-like appearance and it peels off in strips leaving gray or brownish spots on the inner bark [42].IForests 14 01050 i007
Guabisay
Podocarpus sprucei Parl
Tree that grows up to 15 m in height in natural environments. It has a straight trunk and grayish brown fissured outer bark. It has simple, alternate, linear leaves with a hard-textured pointed apex [42].NForests 14 01050 i008
Guaylo
Delostoma integrifolium D. Don
Tree that reaches up to 6 m in height with a cylindrical trunk and smooth bark. It has simple, opposite elliptic to oblong elliptic leaves with entire margin and terminal inflorescence with a few clustered flowers [42].NForests 14 01050 i009
Huesito
Pittosporum undulatum Vent.
Tree from 3 to 5 m in height with grayish brown granular outer bark texture. It has a dense or semi-dense crown of globose or ellipsoidal shape that is light green in color [45].IForests 14 01050 i010
Ramo de novia
Yucca gigantea Lem.
Tree from 3 to 6 m in height; when mature, it usually develops several stems. Its trunk has bulges at the base that taper towards the middle part with grayish brown rough outer bark [46].IForests 14 01050 i011
Sauce
Salix humboldtiana
Tree that reaches between 5 and 12 m in height and 50 cm in diameter. It has a tortuous trunk with cracked outer bark and a wide irregular crown with alternate branching [42].IForests 14 01050 i012
Tilo o Sauco blanco
Sambucus mexicana C. Presl ex DC
Deciduous shrub that reaches up to 3 m or more in height. Its leaves are arranged in opposite pairs [47].IForests 14 01050 i013
Urapan
Fraxinus excelsior L.
Tree that grows up to 35 m in height with irregular crown and deciduous foliage. It has opposite, pinnate compound leaves and the leaflets are finely serrated [48].IForests 14 01050 i014
Table 2. Number of images by class and new rotated images.
Table 2. Number of images by class and new rotated images.
ClassOriginal ImagesRotation 90°Rotation 180°Rotation 270°Rotated Images% Images Rotated
Alamo blanco2951515154513%
Capuli34000000%
Cepillo34000000%
Cepillo blanco34000000%
Cipres2712323236920%
Cucarda322186665%
Eucalipto2493130309127%
Guabisay34000000%
Guaylo2891717175115%
Huesito20844444413239%
Ramo de novia17256565616849%
Sauce34000000%
Tilo o sauco blanco34000000%
Urapan34000000%
4186202187185574
Table 3. Accuracy and kappa of model and interpreter.
Table 3. Accuracy and kappa of model and interpreter.
Full TrainingTop Layer Training
ModelInterpreterModelInterpreter
Base Model NameAccKappaAccKappaAccKappaAccKappa
ResNet V2 1010.9120.9050.8010.7850.7500.7310.6780.653
EfficientNet-Lite0.9220.9160.7700.7520.7800.7630.6480.621
Table 4. Confusion matrix and classification report using ResNet V2 101 base model.
Table 4. Confusion matrix and classification report using ResNet V2 101 base model.
Prediction
1234567891011121314PrecisionRecallF1-Score
Real18200010001001010.870.950.91
23740000111300120.730.860.79
30276214010000000.990.880.93
40008400010001000.950.980.97
50000860000000000.931.000.97
60100080003000020.900.930.91
781900005400010040.900.630.74
80000001830101000.940.970.95
90000000086000000.911.000.96
100110000127500150.940.870.90
110001000000850000.990.990.99
120100001000084000.950.980.97
130300010100008010.980.930.95
141101443011010690.820.800.81
(1) Alamo Blanco, (2) Capuli, (3) Cepillo, (4) Cepillo blanco, (5) Cipres, (6) Cucarda, (7) Eucalipto, (8) Guabisay, (9) Guaylo, (10) Huesito, (11) Ramo de novia, (12) Sauce, (13) Tilo o sauco blanco, (14) Urapan.
Table 5. Confusion matrix and classification report using ResNet V2 101 interpreter.
Table 5. Confusion matrix and classification report using ResNet V2 101 interpreter.
Prediction
1234567891011121314PrecisionRecallF1-Score
Real18400010000000010.700.980.82
266400011113000000.650.740.70
345434213442020030.860.500.63
40008300010001100.930.970.95
51000850000000000.880.990.93
62100077105000000.630.900.74
781710215200010130.570.600.59
82000011780001030.890.910.90
95000010080000000.790.930.86
1022001113194910071.000.570.73
110001000000850000.960.990.97
120100000200079040.980.920.95
132710071100006610.970.770.86
14415161018020000390.640.450.53
(1) Alamo Blanco, (2) Capuli, (3) Cepillo, (4) Cepillo blanco, (5) Cipres, (6) Cucarda, (7) Eucalipto, (8) Guabisay, (9) Guaylo, (10) Huesito, (11) Ramo de novia, (12) Sauce, (13) Tilo o sauco blanco, (14) Urapan.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pacheco-Prado, D.; Bravo-López, E.; Ruiz, L.Á. Tree Species Identification in Urban Environments Using TensorFlow Lite and a Transfer Learning Approach. Forests 2023, 14, 1050. https://doi.org/10.3390/f14051050

AMA Style

Pacheco-Prado D, Bravo-López E, Ruiz LÁ. Tree Species Identification in Urban Environments Using TensorFlow Lite and a Transfer Learning Approach. Forests. 2023; 14(5):1050. https://doi.org/10.3390/f14051050

Chicago/Turabian Style

Pacheco-Prado, Diego, Esteban Bravo-López, and Luis Ángel Ruiz. 2023. "Tree Species Identification in Urban Environments Using TensorFlow Lite and a Transfer Learning Approach" Forests 14, no. 5: 1050. https://doi.org/10.3390/f14051050

APA Style

Pacheco-Prado, D., Bravo-López, E., & Ruiz, L. Á. (2023). Tree Species Identification in Urban Environments Using TensorFlow Lite and a Transfer Learning Approach. Forests, 14(5), 1050. https://doi.org/10.3390/f14051050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop