Next Article in Journal
Development of a Portable, Reliable and Low-Cost Electrical Impedance Tomography System Using an Embedded System
Next Article in Special Issue
Data Analytics and Mathematical Modeling for Simulating the Dynamics of COVID-19 Epidemic—A Case Study of India
Previous Article in Journal
A Novel Rail-Network Hardware Simulator for Embedded System Programming
Previous Article in Special Issue
Evolutionary Machine Learning for Optimal Polar-Space Fuzzy Control of Cyber-Physical Mecanum Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel YOLOv3 Algorithm-Based Deep Learning Approach for Waste Segregation: Towards Smart Waste Management

1
Department of Instrumentation and Control Engineering, Dr. B. R. Ambedkar National Institute of Technology Jalandhar, Punjab 144011, India
2
Electronics and Communication Engineering, PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur 482005, India
3
AI Graduate School, Gwangju Institute of Science and Technology (GIST), Gwangju 61005, Korea
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(1), 14; https://doi.org/10.3390/electronics10010014
Submission received: 30 November 2020 / Revised: 14 December 2020 / Accepted: 17 December 2020 / Published: 24 December 2020
(This article belongs to the Special Issue Evolutionary Machine Learning for Nature-Inspired Problem Solving)

Abstract

:
The colossal increase in environmental pollution and degradation, resulting in ecological imbalance, is an eye-catching concern in the contemporary era. Moreover, the proliferation in the development of smart cities across the globe necessitates the emergence of a robust smart waste management system for proper waste segregation based on its biodegradability. The present work investigates a novel approach for waste segregation for its effective recycling and disposal by utilizing a deep learning strategy. The YOLOv3 algorithm has been utilized in the Darknet neural network framework to train a self-made dataset. The network has been trained for 6 object classes (namely: cardboard, glass, metal, paper, plastic and organic waste). Moreover, for comparative assessment, the detection task has also been performed using YOLOv3-tiny to validate the competence of the YOLOv3 algorithm. The experimental results demonstrate that the proposed YOLOv3 methodology yields satisfactory generalization capability for all the classes with a variety of waste items.

1. Introduction

The rapid explosion in industrialization, urbanization, and global population rate is an attention-grabbing concern, pertaining to environmental degradation. With the global population expanding at an alarming rate, there has been terrific degradation of the environment, resulting in its dreadful condition. As per the published report (2019), India annually generates more than 62 million tons (MT) of solid waste, out of which only 43 MT of the waste is collected, 11.9 MT is treated and almost 31 MT is dumped in landfill sites [1]. Attributable to the existing environmental concerns and improper management of waste, the world encounters gargantuan deleterious effects on the economy, public health and, essentially, the environment. This has shifted the above-all focus towards the worldwide progression of smart cities for guaranteeing effective and smart urban waste management. Moreover, the recycling of waste opens the gateway for research and development and introduces waste to the wealthy business model for sustainable development. However, concerns have been raised towards the requisite of segregation of waste based on its biodegradable and non-biodegradable behavior.
Usually, in the Indian context, wastes consist of paper, plastic, rubber, metal, glass, textiles, organics, sanitary products, electricals and electronics, hazardous substances (paint, spray and chemical) and infectious materials (hospital and clinical), which can be broadly classified as biodegradable (BD) and nonbiodegradable (NBD) waste with their respective share of 52% and 48% [2]. Further, according to the recent Indian government reports, the most common things that have been thrown in the garbage are paper, paper boxes, food, and glass [3]. These things constitute more than 99.5% of total garbage collected, which clearly indicates that people throw dry and wet waste away together. Efficacious waste segregation would assist in the proper disposal and recycling of these wastes based on their biodegradability. Thus, the present era dictates the evolution of a smart waste segregation system to allude to the aforementioned causes of ecological ruin. The segregation of waste is, consequently, seeking attention from various researchers and academicians across the globe [4,5].

2. Motivation

The appropriate classification and organization of wastes into various categories (such as recyclable, biodegradable, nonbiodegradable, organic, harmful, etc.) helps in the proper utilization and disposal of wastes. For waste segregation, computer vision may provide cost effective solutions to identify, classify, and separate out the waste from the humungous dumps of garbage and trash. Owing to the unparalleled developments in the arena of computer vision, waste segregation has been made possible via the identification and detection of wastes through the effective utilization of various approaches for object detection. Usually, the drawbacks of classical object detection techniques (based on Haar cascade classifier, SVMs (Support Vector Machines) or Sliding Window methods) are overcome by deep learning models, for instance, deep convolutional neural networks (DCNNs). Attributable to the ground-breaking achievements in object detection and image classification, researchers across the globe are focusing on the advancements in deep learning approaches for difficult object detection tasks [6,7,8,9,10]. Various deep learning techniques (including CNN, i.e., Convolutional Neural Network [11], R-CNN (Region-based Convolutional Neural Network) family, SSD (Single-shot detector) family [12,13], YOLO (You Only Look Once) [14], etc.) and other classification techniques (SVM and, MLP, i.e., Multi-layer perceptron) [15] have demonstrated their efficiency in performing complex and sophisticated object detection tasks [16,17]. In the context of waste segregation, researchers have shown the competencies of various deep learning strategies in the detection of garbage and its classification [4,5,18,19]. These approaches include CNN and MLP [20], and Faster R-CNN. However, owing to the slow computational capability and slow response time of these approaches, on account of their pipeline execution framework [21], YOLOv3, an improved version of the YOLO [22] algorithm, based on CNNs, emerges to be one of the proficient and promising deep learning algorithms for object detection tasks. The YOLOv3 algorithm yields exceptional results in terms of real-time response rate. Hence, YOLOv3 has established its supremacy over other existing algorithms, along with the members of YOLO family (YOLO, YOLOv2, YOLOv2-tiny and YOLOv3-tiny). Additionally, the efficiency of YOLOv3 in waste segregation is still unexplored. Furthermore, the YOLOv3-tiny Algorithm has been used for various object-detection tasks [23]. Even though the speed of detection is improved by YOLOv3-tiny, some of the detection accuracy is lost. Under the umbrella of the above framework, the present work proposes a maiden attempt towards waste segregation via the effective demonstration of the application of the YOLOv3 algorithm. The efficient capability of the YOLOv3 algorithm in precise object detection with proximate real-time performance has been endorsed in the present investigation. Furthermore, the effectiveness of the YOLOv3 algorithm has been validated by comparing its performance with the YOLOv3-tiny algorithm.
Conclusively, the objective is to replace the waste segregation process, involving human labor by automated waste management processes, wherein the waste will be classified by deep learning approaches. To sum up, the main goal is to be able to successfully identify and classify waste items based on their broad degradability in a reliable manner. To achieve this goal, the work presented in this paper makes the following key contributions:
  • The main contribution of the present investigation is to endorse the efficiency of machine learning and/or deep learning techniques (particularly, YOLO family) for waste segregation based on the broad biodegradable properties of garbage, as these techniques have never been incorporated in this regard, to the best knowledge and belief of authors.
  • The other subcontribution of the present investigation is the development of a garbage image dataset, consisting of 6437 images and distributed among six classes (cardboard, glass, plastic, paper, metal, and organic waste), usually visible in household garbage.
The remainder of this paper is organized as follows. Section 3 describes the dataset utilized in the present investigation. A brief sketch of the YOLOv3 algorithm is presented in Section 4. Then, details of the system specifications and parametric settings used to train the model are presented in Section 5. In Section 6, the experimental results are presented and discussed. To end, Section 7 gives the concluding remarks of the present work.

3. Dataset

The present investigation emphasizes urban wastes in the vicinity of public areas which are frequently disposed of by commuters, pedestrians and occasionally during commercial events. Here, we examine a number of waste items commonly encountered in the surroundings, including BD and NBD items. However, since this was the very first attempt to segregate these items based on the biodegradable property of the material, therefore, a garbage dataset consisting of the most commonly seen waste items need to be developed. For this purpose, 7826 images were acquired in JPEG format using the camera of an Apple iPhone XR (64GB) with 1280 × 960 pixel resolution. After the preprocessing and cleaning of the collected data, 6437 (82%) images were utilized to form a self-made real time dataset wherein each image was labeled with the name of the class to which it belonged and its type (BD/NBD). In the present investigation, these cleaned sample images were grouped into six classes, namely cardboard, glass, plastic, paper, metal and organic waste, as illustrated in Table 1. Furthermore, a detailed description of the waste items assigned to these defined classes, along with their volume size, is provided in Table 2, and their class distribution is represented in Figure 1.
Further, in general, most of the waste items presented in Table 2 belong to only one class. However, it has been observed that some captured images of a few items, such as pizza box, have been made by combining a very thin plastic coating and cardboard, but to efficiently differentiate between these with- and without-plastic coating boxes even with human eyes is very challenging. Moreover, the incorporation of these special cases significantly enhances the complexity and computational overhead. Therefore, in the present analysis, these objects have been considered as objects of the parent class. Another issue that arises with the development of the present garbage dataset is that, in some of the acquired images, one class itself contains one or many other classes; for example, polybags which may contain other class items within them. However, because of the limitation of visual object recognition, these kinds of issues have been resolved by considering only the visible object and classifying accordingly.

4. Methodology: YOLOv3 Algorithm

YOLO (You Only Look Once) is one of the most prominent state-of-the-art deep learning techniques [22] which enables simultaneous object detection and classification. To accomplish the object-detection task, earlier techniques (R-CNN and its variations) employed a pipeline execution architecture, which involves multiple steps. Due to the pipeline architecture and the necessity of the separate training of each individual component, slow speed is achieved along with the increased complexity in optimization. These drawbacks are overcome by YOLO, which transforms object detection into a single regression problem. This performs the simultaneous prediction of multiple bounding boxes and their class probabilities. Unlike sliding window and region proposal-based techniques, the training in YOLO is carried out on full images, thereby directly optimizing the detection performance. However, the real-time speed, end-to-end training capability, along with high average precision and generalization capability of YOLOv3 substantiates its efficiency in performing complex object detection tasks, including significantly small objects.
In general, the YOLOv3 algorithm (as illustrated in Figure 2) simply takes an input image, passes it through a neural network (similar to CNN) to produce an output vector of bounding boxes and class predictions. YOLOv3 extracts a single image, which is then resized to 416 × 416 , that serves as the input to the YOLOv3 neural network. The architecture of the YOLOv3 neural network employed in the darknet-53 framework is illustrated in Figure 3. It consists of convolutional layers, residual layers, upsampling layers, and skip (shortcut) connections. The comprehensive details about the architecture of YOLOv3 are available in an extensive volume of literature works [24].
The YOLOv3 neural network takes an input image to return an output vector (Figure 4). The output vector consists of the following parameters:
  • Prediction Probability (Pc): A probability that each bounding box contains a detectable object.
  • Bounding box properties: Width (Bw), Height (Bh) and Cartesian position (x and y) of the box inside the image (Bx, By).
  • Class probabilities (C1, C2, C3, C4, C5, C6): Probabilities that each object within its bounding box is associated with a specific class.
In YOLOv3, the prediction of bounding boxes is carried out by utilizing the dimension clusters as anchor boxes. Four coordinates (tx, ty, tw, th) are predicted for each bounding box by the YOLOv3 neural network. From the top left corner of the image, in this case, the cell is offset by (Cx, Cy) and the width and height of the bounding box prior are (Pw, Ph), then the corresponding predictions are expressed as Bx, By, Bw and Bh, respectively, as demonstrated in Figure 4. Moreover, if t ^ denotes the ground truth corresponding to a certain coordinate prediction, then the difference of the ground truth value (calculated via the ground truth box) and the estimated prediction (i.e., t ^ t ) is the gradient. By inverting the equations mentioned in Figure 5, the ground truth value can be calculated. In YOLOv3, the score of an object for each bounding box is predicted by utilizing logistic regression. The score of the object is 1 if the overlapping of the bounding-box prior is greatest among all other bounding-box priors w.r.t. the ground truth object. The bounding-box priors, other than the best one, are ignored from prediction, even if their overlapping is greater than the threshold (0.5, in this case). Only one bounding box is assigned for each ground truth object in YOLOv3.

4.1. Performance Parameter Indices

The present investigation examines some of the fundamental key values [24] throughout the training phase to investigate the performance of YOLOv3 in waste segregation. These fundamental key values are as follows.

4.1.1. Precision

Precision is defined in terms of the ratio of the number of objects detected correctly to the number of total objects detected. Mathematically, precision can be computed as expressed by Equation (1).
Precision = N T P N T P + N F P

4.1.2. Recall

Recall is evaluated in terms of the percentage of the number of objects which are correctly detected to the number of ground truth objects. Recall can be evaluated using Equation (2):
Recall = N T P N T P + N F N
where NTP = Number of True Positives, i.e., number of objects detected correctly;
NFP = Number of False Positives, i.e., number of detected objects which could not correspond to the ground truth objects;
NFN = Number of False Negatives, i.e., number of ground truth objects that could not be detected.

4.1.3. Intersection Over Union (IoU)

IoU is a well-known evaluation metric in object detection tasks, which is mathematically represented by Equation (3) and illustrated in Figure 6.
IoU = A B A B
Here, A and B represent the bounding boxes of prediction and ground truth, respectively.

4.1.4. Average Precision (AP)

For a specified threshold value of IoU, a precision–recall curve can be drawn after the identification of the values of precision and recall. The area under the precision–recall curve is referred to as the Average Precision (AP), which can be expressed by Equation (4):
AP = 0 1 p ( r ) d r

4.1.5. Mean Average Precision (mAP)

This signifies the mean of average precisions of all classes defined in the test model and is expressed by Equation (5) for N number of classes.
mAP = i = 1 N AP i N

4.1.6. Loss Function

In the course of training, the sum of squared error loss [22] is used. The computation of the value of the loss function is one of the important criteria in evaluating the performance of YOLOv3 on the test model. Usually, the loss function is defined by Equation (6).
Loss = E r r o r c o o r d + E r r o r I o U + E r r o r c l s
Here, E r r o r c o o r d is the coordinate prediction error, which is expressed by Equation (7).
E r r o r c o o r d = λ c o o r d i = 1 S 2 j = 1 B 1 i j o b j [ { ( x i x i ^ ) 2 + ( y i y i ^ ) 2 } + { ( w i w i ^ ) 2 + ( h i h i ^ ) 2 } ]
Here, ( x i ,   ^ y i ,   ^ w i ,   ^   h i   ^ ) denote the values of the coordinate position, width and height of the predicted bounding box, respectively, and x i , y i , w i and h i signify the true or actual values. Further, λ c o o r d , S 2 , and B represents coordinate error weight, number of grids in the input image, and number of bounding boxes generated by each grid, respectively. The value of 1 i j o b j will be 1 if the object falls into the jth bounding box in grid i. In Equation (6), E r r o r I o U refers to the IoU error expressed by Equation (8).
E r r o r c o o r d = i = 1 S 2 j = 1 B 1 i j o b j ( C i C i ^ ) 2 + λ n o o b j i = 1 S 2 j = 1 B 1 i j o b j ( C i C i ^ ) 2
where λ n o o b j , C i , and C i ^ represent IoU error weight and predicted and true confidence, respectively.
Additionally, E r r o r c l s denotes the classification error and is usually expressed by Equation (9). The E r r o r c l s corresponding to the ith grid is the addition of classification errors associated with all the objects within that grid.
E r r o r c l s = i = 1 S 2 j = 1 B 1 i j o b j c c l a s s e s ( p i ( c ) p i ^ ( c ) ) 2
The notation utilized in Equation (9) is described as below.
cThe specified class to which the detected object belongs;
p true probability that the object belonging to class c is in grid i;
p i ^ predicted value.
In the YOLOv3 algorithm, the input image is distributed into a grid of cells of dimensions S × S , where each grid cell can predict three bounding boxes. Usually, YOLOv3 predicts the bounding boxes at three different scales. For determining the bounding box priors, the k-means clustering is used. In the present investigation, nine clusters and three scales were selected. Further, these clusters were evenly divided across scales and they were distributed as ( 27 × 34 ) ,   ( 56 × 66 ) ,   ( 118 × 177 ) ,   ( 132 × 332 ) ,   ( 225 × 234 ) ,   ( 220 × 354 ) ,   ( 349 × 285 ) ,   ( 302 × 356 ) ,   ( 376 × 367 ) .

4.2. Anchor Boxes

In the case that the midpoints of multiple objects fall in the same grid cell, then the detection of those multiple objects becomes impossible. To avoid this issue, each object in the same grid was assigned with an anchor box. For instance, if we take three anchor boxes, then three predictions can be associated in a single grid cell. Each object is assigned to that anchor box, which has the highest IoU (Intersection over Union). If the IoU is less than the threshold value (here, set to 0.5), then that particular object will not be considered for detection. Thus, the detection of multiple objects in a single grid cell becomes possible by using the sole idea of anchor boxes, as illustrated in Figure 7.

4.3. Non-Max Suppression (NMS)

Another problem encountered in object detection is the multiple detections of the same object, rather than detecting an object just once. The one-time detection of an object is feasible by using non-max suppression. The NMS algorithm compares the bounding box (with max Pc) with all other bounding boxes intersecting with it in a sequential manner. All the bounding boxes associated with the object with comparatively low Pc are suppressed, as demonstrated in Figure 8.

5. The Training

In this work, the entire experimental platform configuration utilized for the training and evaluation of the YOLOv3 neural network are presented in Table 3.
From the cleaned dataset of 6437 images, 80% of the images (5150 images) were used for training purposes and the remaining 20% (1287) were used for testing and validation. Usually, the performance of any deep learning model is highly influenced by the size of the dataset. Generally, training with a small dataset leads to overfitting and to handle such a problem, a transfer learning approach is used [25]. In this approach, a pretrained model is repurposed to accomplish similar detection tasks. This initiates with the training of a base network on a base dataset and task, and then the learned features are transferred to a second target network to be trained on a target dataset and task. This process will tend to work if the features are suitable for both base and target tasks, instead of being specific to the base task. Considering the performance of the YOLO algorithms being trained on the large-scale image dataset (COCO), this work transfers the pre-trained YOLOv3 and YOLOv3-tiny networks on COCO.
At the start of training, the initialization of weights was performed using pretrained weights for the convolutional layers darknet53.conv.74. As discussed earlier, the YOLOv3 neural network is trained for the detection of waste via six classes of objects (cardboard, glass, metal, paper, plastic and organic waste), as discussed in Table 2. Moreover, the performance of the YOLOv3 algorithm has been compared with that of YOLOv3-tiny. For this purpose, the parametric settings used to train the model via the YOLOv3 and YOLOv3-tiny algorithm are tabulated in Table 4. The whole investigation environment uses Visual Studio 2017 for the compilation of the entire script.

6. Performance Evaluation, Results and Discussion

In the present work, the training was accomplished for 12,000 iterations and the total time taken to complete the training simulation was about 48 h on the mentioned platform (illustrated in Table 3) for the YOLOv3 algorithm. During the training simulation, the abovementioned performance parameter indices (including the AP of each class, recall, mAP and average IOU) were examined on a regular time interval. Table 5 presents the results obtained during the training phase with these performance parameter indices for YOLOv3 and YOLOv3-tiny. As evident from Table 5, after the 5000th iteration, mAP reaches around 94% by YOLOv3; however, YOLOv3-tiny yields only 45.96%. Thereafter, the mAP value settled down at approximately the same value (94.99%, best value) for YOLOv3. However, the mAP for the YOLOv3-tiny algorithm for the present experiment does not settle even after 12,000 iterations and attains the best value of 51.95%.
The variations in average loss and mAP values w.r.t the number of iterations during training by YOLOv3 and YOLOv3-tiny algorithm are illustrated in Figure 9a,b, respectively. Additionally, for comparative analysis, training was also carried out using the YOLOv3-tiny algorithm on the same dataset, which took approximately 14 h on a similar system configuration, as shown in Table 3. As is evident from the results illustrated in Figure 9a,b, the average loss function value using YOLOv3 and YOLOv3-tiny are 0.6806 and 0.1525, respectively, after the completion of training (12,000 iterations). Conclusively, our simulation experimental training results indicate that the mAP value for YOLOv3 is 82.85% higher than that of the YOLOv3-tiny model, in reference with the best value, which strengthens and validates our hypothesis.
From Figure 10, it is observed that YOLOv3 offers enhanced AP for each class during experimental training simulation. The trends obtained during the training for all the mentioned classes (Table 2) indicate that YOLOv3 is, again, demonstrating outstanding performance as compared to YOLOv3-tiny in terms of AP, as illustrated in Figure 10. For instance, the best AP value (as illustrated in Table 5) attained by YOLOv3 for various classes (cardboard, glass, plastic, paper, metal and organic waste) are 97.27%, 97.40%, 99.87%, 85.28%, 91.16% and 98.93%, respectively; however, YOLOv3-tiny attains 62.16%, 61.79%, 31.98%, 48.32%, 26.15% and 81.29%, respectively, for the same classes. To provide a comparative insight into the performance of these two approaches in terms of mAP (%), a comparative sketch is provided, illustrating the variations in mAP values with the increasing number of iterations during training by YOLOv3 and YOLOv3-tiny. This comparison is demonstrated in the form of chart, as portrayed in Figure 11.
Furthermore, a statistical analysis of the present work in terms of AP, mAP and detection speed (frames per second, i.e., FPS) is presented in Table 6. The data reveal the effectiveness of YOLOv3 over the YOLOv3-tiny algorithm.
After simulating the training, the test images were validated on the trained model. The test images correspond to the image garbage test set developed in this paper, which has 1287 images, including 165 cardboard, 163 glass, 146 metal, 312 paper, 317 plastic and 184 organic waste samples. The obtained experimental results demonstrate that the detection capability and prediction probability of YOLOv3 is significantly higher than YOLOv3-tiny, as visualized in Figure 12 and presented in Table 7. Most of the test images were accurately detected with acceptable prediction probability by YOLOv3. YOLOv3 gives true predictions of all the test images and is capable of detecting even small-size objects. However, YOLOv3-tiny gives false predictions for test images 1, 2, 6, 8 and 9. In addition, YOLOv3-tiny does not predicts test image 7 at all.
Table 8 presents the missed and false detection rates for both algorithms. Evidently, these detection rates are comparatively low (in the case of YOLOv3) when compared to the YOLOv3-tiny algorithm. Furthermore, the test results illustrate that the simulation time in the prediction to classify the objects using YOLOv3-tiny is lowered by approximately four-times more than YOLOv3, as shown in Table 7. This means that the computation speed of YOLOv3-tiny is significantly higher than YOLOv3. Conclusively, although the computational performance of YOLOv3-tiny is remarkable, the detection capability and prediction probabilities are not acceptable.
Furthermore, to quantify the obtained results, a comparison of detection capability in test images has been made among the models developed by employing YOLOv3 and YOLOv3-tiny. As depicted in Table 9, the trained YOLOv3 model achieved superior detection capability compared to YOLOv3-tiny. It was able to detect most of the objects in the test images with significant prediction probability. Specifically, YOLOv3 achieved 100% detection accuracy for test images 2, 3, 4, 5, 6, and 7; however, YOLOv3-tiny achieved this for very simple test images (3, 4, and 5). Furthermore, the average accuracy for all the test images was been obtained as 85.29% and 26.47% for YOLOv3 and YOLOv3-tiny, respectively. Therefore, YOLOv3 dominates YOLOv3-tiny by a notable margin of 58.82%. However, YOLOv3 also struggles in the accurate detection of objects, particularly under occlusion and complex environmental conditions (test images 1, 8, and 9), as illustrated in Figure 12 and Table 9. This might be because of very small visual appearances and cluttered backgrounds.

7. Conclusions

This paper presented a novel application of the YOLOv3 algorithm for waste segregation as an aid to strengthen smart urban waste segregation and management framework. The neural network was trained on a self-made dataset of 6437 images of urban waste products for the detection of six classes of waste items. The obtained experimental and investigational results demonstrate the efficiency of the proposed work in the segregation of waste into two different categories—biodegradable and nonbiodegradable. The near real-time detection of waste was accomplished in this work. The quantitative comparison of the results obtained by YOLOv3 and YOLOv3-tiny endorse the efficacy of YOLOv3 in waste segregation. Furthermore, the improved prediction probability by YOLOv3 demonstrates its effectiveness over YOLOv3-tiny. Conclusively, the comparative analysis between YOLOv3-tiny and YOLOv3 quantified the percentage improvement in speed with reduced accuracy (due to the simplified architecture of YOLOv3-tiny) which, in turn, helped in understanding the accuracy–speed trade-off. Furthermore, the garbage image detection process consisted of many complexities, such as objects made up of more than one type of material and may inherited other-class objects. To deal such real time complexities, objects belonging to parent classes and objects of visible category only were considered; however, this opens the window for further research to more exactly classify garbage, depending upon the property of material(s). Additionally, the object detection strategy for waste segregation utilized in this work opens the gateway for the effective recycling and disposal of waste. However, the reduction in the time of detection along with exceptionally high prediction probability provides scope for further research. Future work will focus on the optimization of results, along with the prediction probability for other waste items in the real world.

Author Contributions

Conceptualization, S.K. and O.P.V.; methodology, S.K. and H.G.; software, S.K., O.P.V. and I.A.A.; validation, S.K., O.P.V. and I.A.A.; formal analysis, H.G., D.Y. and C.W.A.; investigation, S.K., D.Y. and O.P.V.; data curation, S.K., D.Y. and H.G.; writing—original draft preparation, S.K., D.Y. and H.G.; writing—review and editing, O.P.V., I.A.A. and C.W.A.; visualization, H.G. and I.A.A.; supervision, O.P.V. and C.W.A.; funding acquisition, C.W.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (MSIT) (No. NRF-2019R1I1A2A01057603).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. India’s Trash Bomb: 80% of 1.5 Lakh Metric Tonne Daily Garbage Remains Exposed, Untreated—India News. Available online: https://www.indiatoday.in/india/story/india-s-trash-bomb-80-of-1-5-lakh-metric-tonne-daily-garbage-remains-exposed-untreated-1571769-2019-07-21 (accessed on 20 November 2020).
  2. Sharma, K.D.; Jain, S. Overview of Municipal Solid Waste Generation, Composition, and Management in India. J. Environ. Eng. 2019, 145, 04018143. [Google Scholar] [CrossRef]
  3. SURAT KHUB SURAT-1. Available online: https://darpg.gov.in/sites/default/files/19.DoortoDoorGarbageCollectionsystem.pdf (accessed on 13 December 2020).
  4. Wang, Y.; Zhang, X. Autonomous garbage detection for intelligent urban management. MATEC Web Conf. 2018, 232, 01056. [Google Scholar] [CrossRef] [Green Version]
  5. Devi, R.S.S.; Vijaykumar, V.R.; Muthumeena, M. Waste segregation using deep learning algorithm. Int. J. Innov. Technol. Explor. Eng. 2018, 8, 401–403. [Google Scholar]
  6. Lu, H.; Zhang, M.; Xu, X.; Li, Y.; Shen, H.T. Deep Fuzzy Hashing Network for Efficient Image Retrieval. IEEE Trans. Fuzzy Syst. 2020. [Google Scholar] [CrossRef]
  7. Lu, H.; Li, Y.; Chen, M.; Kim, H.; Serikawa, S. Brain Intelligence: Go beyond Artificial Intelligence. Mob. Netw. Appl. 2018, 23, 368–375. [Google Scholar] [CrossRef] [Green Version]
  8. Lu, H.; Li, Y.; Mu, S.; Wang, D.; Kim, H.; Serikawa, S. Motor anomaly detection for unmanned aerial vehicles using reinforcement learning. IEEE Internet Things J. 2018, 5, 2315–2322. [Google Scholar] [CrossRef]
  9. Chen, Z.; Lu, H.; Tian, S.; Qiu, J.; Kamiya, T.; Serikawa, S.; Xu, L. Construction of a Hierarchical Feature Enhancement Network and Its Application in Fault Recognition. IEEE Trans. Ind. Inform. 2020. [Google Scholar] [CrossRef]
  10. Huang, R.; Gu, J.; Sun, X.; Hou, Y.; Uddin, S. A Rapid Recognition Method for Electronic Components Based on the Improved YOLO-V3 Network. Electronics 2019, 8, 825. [Google Scholar] [CrossRef] [Green Version]
  11. Dundar, A.; Jin, J.; Martini, B.; Culurciello, E. Embedded streaming deep neural networks accelerator with applications. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 1572–1583. [Google Scholar] [CrossRef] [PubMed]
  12. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2016; Volume 9905 LNCS, pp. 21–37. [Google Scholar]
  13. Fu, C.-Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. DSSD: Deconvolutional Single Shot Detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  14. Lu, S.; Wang, B.; Wang, H.; Chen, L.; Linjian, M.; Zhang, X. A real-time object detection algorithm for video. Comput. Electr. Eng. 2019, 77, 398–408. [Google Scholar] [CrossRef]
  15. Dollár, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian detection: An evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 743–761. [Google Scholar] [CrossRef] [PubMed]
  16. Pathak, A.R.; Pandey, M.; Rautaray, S. Application of Deep Learning for Object Detection. In Proceedings of the Procedia Computer Science; Elsevier B.V.: Amsterdam, The Netherlands, 2018; Volume 132, pp. 1706–1717. [Google Scholar]
  17. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object Detection with Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Adedeji, O.; Wang, Z. Intelligent waste classification system using deep learning convolutional neural network. In Proceedings of the Procedia Manufacturing; Elsevier B.V.: Amsterdam, The Netherlands, 2019; Volume 35, pp. 607–612. [Google Scholar]
  19. Chu, Y.; Huang, C.; Xie, X.; Tan, B.; Kamal, S.; Xiong, X. Multilayer hybrid deep-learning method for waste classification and recycling. Comput. Intell. Neurosci. 2018, 2018, 5060857. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Sung, K.K.; Poggio, T. Example-based learning for view-based human face detection. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 39–51. [Google Scholar] [CrossRef] [Green Version]
  21. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; Volume 2016-Decem, pp. 779–788. [Google Scholar]
  22. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  23. Nordin, P.; Lidström, F. Object Detection Using Yolov3 Tiny. 2019. Available online: http://urn.kb.se/resolve?urn=urn:nbn:se:oru:diva-76453 (accessed on 20 November 2020).
  24. Nie, X.; Yang, M.; Liu, R.W. Deep Neural Network-Based Robust Ship Detection Under Different Weather Conditions. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019, Auckland, New Zealand, 27–30 October 2019. [Google Scholar]
  25. Alom, Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, S.; Hasan, M.; Van Essen, B.C.; Awwal, A.A.S.; Asari, V.K. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics 2019, 8, 292. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Statistical analysis of class distribution of dataset.
Figure 1. Statistical analysis of class distribution of dataset.
Electronics 10 00014 g001
Figure 2. Outline of YOLOv3 algorithm.
Figure 2. Outline of YOLOv3 algorithm.
Electronics 10 00014 g002
Figure 3. YOLOv3 architecture.
Figure 3. YOLOv3 architecture.
Electronics 10 00014 g003
Figure 4. Prediction of Bounding Box.
Figure 4. Prediction of Bounding Box.
Electronics 10 00014 g004
Figure 5. Location prediction via bounding box.
Figure 5. Location prediction via bounding box.
Electronics 10 00014 g005
Figure 6. Illustration of the concept of IoU (Intersection Over Union).
Figure 6. Illustration of the concept of IoU (Intersection Over Union).
Electronics 10 00014 g006
Figure 7. Illustration of anchor boxes.
Figure 7. Illustration of anchor boxes.
Electronics 10 00014 g007
Figure 8. Non-max suppression for filtering multiple detections.
Figure 8. Non-max suppression for filtering multiple detections.
Electronics 10 00014 g008
Figure 9. Variations of loss function value and mAP value w.r.t. number of iterations for (a) YOLOv3 (b) YOLOv3-tiny.
Figure 9. Variations of loss function value and mAP value w.r.t. number of iterations for (a) YOLOv3 (b) YOLOv3-tiny.
Electronics 10 00014 g009
Figure 10. Variation in AP w.r.t. number of iterations for various classes via YOLOv3 and YOLOv3-tiny algorithm (a) cardboard (BD) (b) glass (NBD) (c) metal (NBD) (d) paper (BD) (e) plastic (NBD) (f) organic waste (BD).
Figure 10. Variation in AP w.r.t. number of iterations for various classes via YOLOv3 and YOLOv3-tiny algorithm (a) cardboard (BD) (b) glass (NBD) (c) metal (NBD) (d) paper (BD) (e) plastic (NBD) (f) organic waste (BD).
Electronics 10 00014 g010
Figure 11. Variation in mAP w.r.t. number of iterations via YOLOv3 and YOLOv3-tiny algorithm.
Figure 11. Variation in mAP w.r.t. number of iterations via YOLOv3 and YOLOv3-tiny algorithm.
Electronics 10 00014 g011
Figure 12. Experimental results of object detection for waste segregation on various test images (a) original image (b) detection by YOLOv3-tiny (c) detection by YOLOv3.
Figure 12. Experimental results of object detection for waste segregation on various test images (a) original image (b) detection by YOLOv3-tiny (c) detection by YOLOv3.
Electronics 10 00014 g012aElectronics 10 00014 g012bElectronics 10 00014 g012c
Table 1. Illustration of sample images with their respective class.
Table 1. Illustration of sample images with their respective class.
Electronics 10 00014 i001 Electronics 10 00014 i002 Electronics 10 00014 i003 Electronics 10 00014 i004 Electronics 10 00014 i005 Electronics 10 00014 i006
(a) Class 1: Cardboard(b) Class 2: Glass(c) Class 3: Metal(d) Class 4: Paper(e) Class 5: Plastic(f) Class 6: Organic waste
Table 2. Description and statistics of Garbage Images in the dataset.
Table 2. Description and statistics of Garbage Images in the dataset.
Class (Type)ItemsQuantity
Cardboard (BD)Pizza boxMailing boxPaperboard box825
Gift boxPacking box
Glass (NBD)Glass JarGlass cupMirror816
Soft drink bottleWine BottleWindow Glass
Metal (NBD)Soft Drink CansBeer CansBlades730
Water Bottles
Paper (BD)NewspapersPaper GlassesNotebooks1561
TeacupsBooksPosters
Plastic (NBD)Water BottlesMilk BottlesPolybags1583
Plastic Jars
Organic Waste (BD)FruitsVegetables 922
BD: Biodegradable, NBD: Non-Biodegradable.
Table 3. Experimental platform configuration.
Table 3. Experimental platform configuration.
SpecificationDetails
Operating SystemWindows, 64-bit Operating System
CPUIntel(R) Core(TM) i7-9700F CPU @ 3.00 GHz
RAM8 GB
GPUMSI Gaming GeForce GTX 1650 Super 128-Bit HDMI/DP/DVI 4GB GDRR6 HDCP Support DirectX 12 Dual Fan VR Ready OC Graphics Card
GPU acceleration libraryCUDA10.0, CUDNN7.4
Table 4. Parameters of CFG used for training our model.
Table 4. Parameters of CFG used for training our model.
ParameterValue(s)
YOLOv3 Neural NetworkYOLOv3-tiny Neural Network
Width416416
Height416416
Batch6464
Subdivisions *6416
Channels3-
Momentum0.90.9
Decay0.00050.0005
Learning rate0.0010.001
Maximum number of Batches *12,00012,000
PolicyStepsSteps
Steps *4800, 54004000, 4500
Scale0.1, 0.10.1, 0.1
Classes *66
Filters * ( 4 + 1 + 6 ) × 3 = 33 ( 4 + 1 + 6 ) × 3 = 33
* Represents the parameters modified in the original YOLOv3 CFG and YOLOv3-tiny CFG, respectively. Note: Filters usually depend on the number of classes, bounding box properties, prediction probability, and the number of masks, i.e., filters = {number of bounding box properties (4) + Prediction probability Pc (1) + Total number of classes (6)} × Number of masks, where mask denotes the indices of anchors (3).
Table 5. Comparative study of simulation training results of YOLOv3 with YOLOv3-tiny on the test model.
Table 5. Comparative study of simulation training results of YOLOv3 with YOLOv3-tiny on the test model.
IterationsYOLO VersionClasses AP (%)mAP (%)RecallAverage IoU (%)
Cardboard (BD)Glass (NBD)Metal (NBD)Paper (BD)Plastic (NBD)Organic Waste (BD)
1000v331.6029.5740.1915.8610.0357.5830.800.2121.63
v3-tiny16.5416.9617.0623.686.7715.6716.110.1314.81
2000v379.3180.7488.6058.2771.0991.0578.160.6846.11
v3-tiny33.5351.3915.4021.9733.8221.3729.580.3324.18
3000v391.0190.5496.9380.1482.1397.3189.670.8156.54
v3-tiny60.4350.8839.2838.9821.3757.8744.770.3247.50
4000v391.0687.6799.5675.3486.7596.0089.400.8647.45
v3-tiny34.8139.8516.9322.616.7239.8726.400.3222.33
5000v395.7596.9299.7984.2190.4498.7694.310.8170.20
v3-tiny51.9246.3632.1948.0226.8970.4545.960.4447.08
6000v393.8394.3599.6081.6189.9798.4892.980.8955.21
v3-tiny58.2852.2334.1245.3023.1374.1647.870.4746.04
7000v394.8597.0399.7584.7989.5599.0694.170.7771.45
v3-tiny62.2254.9027.8343.8919.1469.6746.280.4940.95
8000v394.5996.5599.6984.9887.5599.3993.790.7170.88
v3-tiny56.6056.1429.5245.0020.3773.4946.850.5043.36
9000v395.0692.1199.7182.3489.9197.8392.830.8954.76
v3-tiny47.0650.0423.2840.7214.7060.1339.320.4335.46
10,000v395.7794.9299.7984.3891.1598.5094.080.9059.86
v3-tiny53.5553.8131.7742.7118.1071.5745.250.4842.61
11,000v396.8297.4399.8984.6290.4398.2994.580.8865.69
v3-tiny54.9953.5330.3642.4418.0971.3645.130.4741.54
12,000v395.6495.9699.8484.9591.1998.4994.350.9061.79
v3-tiny42.6052.8847.8849.1627.6971.0848.550.4750.35
BEST *v397.2797.4099.8785.2891.1698.9394.990.8767.42
v3-tiny62.1661.7931.9848.3226.1581.2951.950.5247.20
* BEST represents the iteration for which maximum mAP has been observed during training. AP: Average Precision. mAP: Mean Average Precision.
Table 6. Comparison of YOLOv3 and YOLOv3-tiny algorithms in terms of average precision and detection speed.
Table 6. Comparison of YOLOv3 and YOLOv3-tiny algorithms in terms of average precision and detection speed.
AlgorithmClasses AP (%)mAP
(%)
Detection Speed (FPS)
Cardboard (BD)Glass (NBD)Metal (NBD)Paper (BD)Plastic (NBD)Organic Waste (BD)CPU *On Specified Platform (Table 3)
YOLOv397.2797.4099.8785.2891.1698.9394.990.33.0
YOLOv3-tiny62.1661.7931.9848.3226.1581.2951.952.921.7
IOU = 0.75; mAP denotes the mean AP. * CPU represents the detection on Intel(R) Core(TM) i5-4200U CPU @ 1.60GHz 2.30GHz.
Table 7. Quantitative comparison of the Experimental results of the test model.
Table 7. Quantitative comparison of the Experimental results of the test model.
Test ImageItemsPrediction Probability (in %)Prediction Time (in Milliseconds)
YOLOv3YOLOv3-tinyYOLOv3YOLOv3-tiny
1Plastic (NBD)78False Detection228.3052.13
Plastic (NBD)88
Plastic (NBD)92
Plastic (NBD)55
2Paper (BD)98False Detection248.5051.86
Plastic (NBD)89
3Paper (BD)10027222.4251.30
4Cardboard (BD)10085222.5751.39
5Organic Waste (BD)9589252.4751.45
6Glass (NBD)94False Detection247.9451.93
Glass (NBD)83
Glass (NBD)93
7Glass (NBD)98No Detection252.6452.00
8Plastic (NBD)61False Detection261.4259.91
Plastic (NBD)73
Plastic (NBD)62
Plastic (NBD)55
Plastic (NBD)88
Plastic (NBD)99
Plastic (NBD)80
Plastic (NBD)78
Plastic (NBD)68
Plastic (NBD)78
Plastic (NBD)79
Plastic (NBD)82
9Plastic (NBD)88False Detection247.9751.98
Paper (BD)68
Paper (BD)72
Paper (BD)92
Table 8. Comparison of missed and false detections with YOLOv3 and YOLOv3-tiny algorithms.
Table 8. Comparison of missed and false detections with YOLOv3 and YOLOv3-tiny algorithms.
AlgorithmMissed Detection Rate (%)False Detection Rate (%)MAP (%)
YOLOv300.5794.99
YOLOv3-tiny4.7811.5851.95
Table 9. Comparison of detection capability.
Table 9. Comparison of detection capability.
Test ImageGround TruthObjects Detected
YOLOv3YOLOv3-tiny
1641
2221
3111
4111
5111
6331
7110
814122
9541
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kumar, S.; Yadav, D.; Gupta, H.; Verma, O.P.; Ansari, I.A.; Ahn, C.W. A Novel YOLOv3 Algorithm-Based Deep Learning Approach for Waste Segregation: Towards Smart Waste Management. Electronics 2021, 10, 14. https://doi.org/10.3390/electronics10010014

AMA Style

Kumar S, Yadav D, Gupta H, Verma OP, Ansari IA, Ahn CW. A Novel YOLOv3 Algorithm-Based Deep Learning Approach for Waste Segregation: Towards Smart Waste Management. Electronics. 2021; 10(1):14. https://doi.org/10.3390/electronics10010014

Chicago/Turabian Style

Kumar, Saurav, Drishti Yadav, Himanshu Gupta, Om Prakash Verma, Irshad Ahmad Ansari, and Chang Wook Ahn. 2021. "A Novel YOLOv3 Algorithm-Based Deep Learning Approach for Waste Segregation: Towards Smart Waste Management" Electronics 10, no. 1: 14. https://doi.org/10.3390/electronics10010014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop