Next Article in Journal
Multi-Sensor Spectral Imaging of Geological Samples: A Data Fusion Approach Using Spatio-Spectral Feature Extraction
Previous Article in Journal
Predicting Group Contribution Behaviour in a Public Goods Game from Face-to-Face Communication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning-Based Automatic Mosquito Sensing and Control System for Urban Mosquito Habitats

1
Department of Biological Sciences, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Daejeon 34141, Korea
2
Department of Civil and Environmental Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Daejeon 34141, Korea
3
Department of Mechanical Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Daejeon 34141, Korea
4
School of Electrical Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2019, 19(12), 2785; https://doi.org/10.3390/s19122785
Submission received: 9 April 2019 / Revised: 16 June 2019 / Accepted: 18 June 2019 / Published: 21 June 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Mosquito control is important as mosquitoes are extremely harmful pests that spread various infectious diseases. In this research, we present the preliminary results of an automated system that detects the presence of mosquitoes via image processing using multiple deep learning networks. The Fully Convolutional Network (FCN) and neural network-based regression demonstrated an accuracy of 84%. Meanwhile, the single image classifier demonstrated an accuracy of only 52%. The overall processing time also decreased from 4.64 to 2.47 s compared to the conventional classifying network. After detection, a larvicide made from toxic protein crystals of the Bacillus thuringiensis serotype israelensis bacteria was injected into static water to stop the proliferation of mosquitoes. This system demonstrates a higher efficiency than hunting adult mosquitos while avoiding damage to other insects.

1. Introduction

Mosquitoes annually cause significant damage to mankind as they disseminate various deadly infectious diseases including malaria, yellow fever, or encephalitis [1]. The recent outbreak of the Zika virus [2] has again awakened the public to the dangers of mosquitoes and the necessity of vector control. Classic mosquito control approaches are generally focused on the control of the adult mosquito via thermal fogging trucks, pesticides, or even automated electrical mosquito traps [3]. However, the mosquito is a small, flying insect which tends to hide near urban areas. Consequently, pesticides and fogs have difficulty in permeating into these hideouts. Given the impact on the surrounding ecosystem and the need to use these preventive measures near cities, the toxicity of the chemicals is also limited. Mosquito nets represent a preventive measure, but not a fundamental solution, for reducing the proliferation of mosquitoes.
Compared to the imago stage, mosquito larvae tend to proliferate and gather in static water such as puddles. This feature makes the control of mosquito larvae much more efficient than controlling them at the imago stage [4]. The well-known habitats of mosquito larvae are the puddles generated after rainfall. However, recent research has shown that most mosquito habitats found in urban areas are located near rainfall collection wells [5] and sewage tanks located under buildings [6]. These urban/indoor facilities enable mosquitoes to increase in number throughout the year, and not only in summer when rain is frequent.
Many larvicides have been developed to kill mosquito larvae, and one of the most promising larvicides is composed of toxic protein crystals extracted from the spores of the Bacillus thuringiensis serotype israelensis (Bti) bacteria [4]. The toxicity is species specific and a low rate of tolerance was observed. Although the Bti treatment is effective and eco-friendly, it does not affect a large area and is washed away rapidly (in less than a month). Furthermore, it is impossible to spray Bti larvicide into every candidate puddle or region where mosquito larvae can grow using human labor alone.
Instead of monitoring every water-related facility in a given urban area, we propose an automated mosquito detection/larvicide spray system installed at the mosquito habitat candidate points. Additionally, instead of passive traps (no detection) or a simple object detector [3], which counts the presence of flying insects, we try to obtain static images of the mosquitoes with a camera for image processing-based mosquito sensing.
Recent advances in deep neural network structures represented by the convolutional neural network (CNN) have demonstrated a remarkable increase in performance compared to conventional machine learning methods [7,8,9]. Many applications based on deep learning have been developed, including vermin monitoring such as jellyfish swarm detection [10]. This research has also applied deep learning based-architecture for the image processing. As the importance of the control of mosquitoes has increased, some research on detecting mosquitoes from a single image using neural networks have been conducted [11,12,13].
The proposed system is equipped with a clean, white mosquito observation pad that lures adult mosquitoes. An attached camera records images of the mosquitoes on the pad, attracted by the lures. If the mosquito larvae proliferate at the installed point (a constant or increasing number of mosquitoes is observed), the attached robot activates and drops a Bti larvicide package into the water to exterminate the growing mosquito larvae. Therefore, our proposed system is a novel approach to controlling mosquitoes, which not only uses deep learning-based image classification, but also uses own hardware equipped with a luring pad and Bti dropper. In addition, the whole control system is placed in an indoor experimental site and verifies that it can exterminate the mosquito larvae.

2. Materials and Methods

2.1. Platform Overview

The system is equipped with a mosquito observation pad and a linked camera for mosquito image acquisition. To lure the mosquitoes, an Ultra Violet (UV) Light-Emitting Diode (LED), a 39 °C heat pad, and CO2-generating chemicals (sodium hydrogen carbonate and acetic acid) were used [14,15]. Fruit flavor was added to the acetic acid for a better luring effect. Mosquitoes tend to land on nearby walls or structures before approaching the target (the CO2 source in this case). The white observation pad was designed to induce mosquito landing and block the background for easier image processing. For the lighting of the observation pad, a white visible light LED lamp equipped with a timer was attached. Streamed images from the webcam are sent to the computer for image processing. The overall structure of the system is shown in Figure 1. The hardware system developed for this purpose is shown in Figure 2.

2.2. Mosquito Luring Experiments

The mosquito luring experiment was conducted in the basement of the W2 building at Korea Advanced Institute of Science and Technology (KAIST). The indoor sewage tank was placed in the basement of the building and the dissemination of the mosquitoes was reported. The developed sensor node was placed in the room near the sewage tank. The sensor node was placed in the dark for 15 min and the lamp was turned on and off in a 15 min cycle. The attached camera recorded the image of the observation pad and was subsequently used for the image-based mosquito sensing. As a negative control of the mosquito lure, an experiment with an empty observation pad without CO2-generating chemicals was also performed for comparison.

2.3. Image Preparation and Deep Learning Architecture

The deep learning architecture-based pipeline was developed for image processing and mosquito detection in the recorded videos. The observation pad attached to the sensor node blocks the background noise and generates a plain background in the video. The grayscale conversion and thresholding were performed to find black objects on the screen. As the distance between the observation pad where the mosquitoes land and the camera is fixed, the size of the mosquitoes generally falls into a certain boundary. The contour in the threshold images was filtered based on their size and saved as 60 × 60 color image patches. The collected patches were manually classified into mosquito and non-mosquito images and the deep learning image classifier was trained to recognize them. AlexNet [7] was originally designed for the ImageNet database [16] classification task with a learning rate of 0.001 and with a step-down decay used for the classification. Of the collected image patches, 25% (approximately 2000 images in total) were used as a test dataset while the rest of the images were used for the training.
Classifying single image patches obtained by color filtering is a feasible design. However, color is not a very robust method even though the observation pads block the background. Also, classifying every dark object in the image requires multiple forward passes of a network, which requires unnecessarily long computing time and resources. Instead, this research applied a fully convolutional network (FCN) [17] and neural network-based regression to count the number of mosquitoes in the image with the least network inference. The summary of the network is shown in Table 1. The total number of parameters in the FCN is approximately 134.82 M.
Label image datasets were generated by changing all pixel values to black (0,0,0 in RGB) except the mosquito regions detected by the pre-trained patch classifying network. The regions corresponding to the mosquitoes were changed to purple (255,0,255 in RGB). The FCN network, which infers the possibility that each pixel belongs to the given classes, was used for detecting the mosquito-like regions in the image. After the training, the probability map before the final deconvolution layer is extracted and passed to the regression network. The same AlexNet structure used in the patch classification was leveraged but the loss layer was changed to the Euclidean distance loss from the softmax cross entropy for regression purposes. The number of mosquitoes in the image was used as a label. Labels of 0 to 5 were prepared and numbers larger than 5 were considered as label 5. Approximately 60 images per label were prepared. The overall pipeline process is shown in Figure 3. A Caffe framework [18] with Python binding was used for the implementation of the system. The summary of the network is shown in Table 2. The total number of parameters in AlexNet is approximately 60.97 M.
The score map before the final deconvolution layer was extracted and used as a final output as shown in Figure 3b. The original purpose of the FCN was to obtain the outline of the target object in the image by examining the pixel-wise associability of the image. However, the FCN was not used in this research to detect the exact edge of the mosquitoes in the image; instead, it marked the rough location of the mosquitoes in the image. By doing so, the number of images required to train the FCN was decreased (the FCN requires a calculated probability over a certain threshold to generate a final output edge and many training samples are required to increase pixel-wise certainty); however, a probability map with rough accuracy is still obtainable.

3. Results and Discussion

3.1. Mosquito Luring Experiment

The experiment results demonstrated that the system successfully attracted mosquitoes during its operation. A recorded video showed that the mosquitoes started to approach the trap approximately 14 min after its installation. The number of mosquitoes present at random moments during the 15 min video was manually calculated for statistical purposes. The image of the lured mosquitoes and a boxplot generated with 10 data points from the 10 trials are shown in Figure 4. More than three mosquitoes constantly (median value of three) approached the trap during its operation with a maximum number of eight. The negative control (no luring mechanism used) showed zero or one mosquito during multiple trials. Other insects were not attracted to the observation pad during the luring experiment.
The system is not equipped with an adult mosquito killing or trapping mechanism. UV/visible light-based mosquito traps attract not only mosquitoes, but also useful pollinating insects, thereby causing collateral damage. Although the attracted insects may cause false positive image processing results, only a larvicide drug package is released to avoid killing unwanted insects. Furthermore, the elimination of active traps requiring actuators can help reduce power consumption.

3.2. Deep Learning-Based Mosquito Detection

The intermediate deep neural networks required for the procedure introduced in Section 2.3 were developed and trained. Firstly, the AlexNet classifier network for the mosquito/non-mosquito contour classification shown in Figure 3a was trained and examined. The trained network showed an accuracy of 95.25% at the test phase, at the 20th epoch after the training started. Training and test loss started from a value of 0.6 and 0.5 and dropped to 0.15 and 0.2 after all iterations were completed.
As the training results demonstrated sufficient accuracy in detecting mosquitoes, the prepared network was used to generate mosquito colored-only label images for the FCN training. Approximately 100 images with the corresponding label images were generated and used for the FCN training. The stride size of 32 was used for faster inference time/low memory consumption compared to the other stride size options (16 and 8). The input image and corresponding FCN processed output is shown in Figure 5. The probability that the pixel falls into a region where a mosquito is present is marked with dark blue, while the other less probable regions are colored with a red to green spectrum. The blue dotted region is similarly located at the position where the mosquito exists in the input image. If the system tries to detect the mosquitoes by color-based contour finding and classification, tuning of the thresholding value is required, as it affects the final accuracy of the classifier. Also, multiple inferences, although required, drop the accuracy of the total result and require a certain amount of computation time for each image. By using the FCN, the proposed system can infer the possible number of mosquitoes with a single network inference. However, the generated probability map does not exactly correspond to the mosquito region. Counting the number of blue dots is not applicable as closely-located mosquitoes fall into the same boundary and are marked as a larger single dot. Thus, the generated score maps were passed to the regression network for neural network-based regression to estimate the number of mosquitoes based on the probability map.

3.3. Examination of the Mosquito Counting Pipeline

The generated score maps were trained to the modified AlexNet introduced in Section 2.3. The number of mosquitoes in the image was used as a label for the regression and ground truth for the performance measurement. As a comparison, the image classification network classifying the contour obtained from the image was compared with the FCN and neural network-based regression. The results are shown in Figure 6.
A total of 80 test images were processed and compared by the two networks. The label is limited to 5 as most images have less than 5 mosquitoes. A few images containing more than 5 mosquitoes were also referred to with a label of 5. If the network output exceeded 5 while processing the label 5 image, it was considered as a correct answer. The FCN and neural network-based regression result demonstrated an accuracy of 84%. Meanwhile, the single image classifier demonstrated an accuracy of only 52%. As shown in Figure 6, the blue line indicating the inference results from the classifier network has more divergence from the ground truth (the red dotted line) compared to the results from the regression network (the green line). As previously mentioned, though the trained classifier shows good classification results when classifying a single image patch, giving perfect answers for every image patch is a difficult task. Root mean square errors (RMSEs) of 1.37 (blue) and 0.42 (green) were obtained, showing that the regression network with the FCN score map input has much better performance in number estimation tasks. The proposed FCN and neural network-based regression demonstrated an error of ±1 most of the time, while the classification network had errors within a much higher range.
Although it is possible to count each mosquito, it is necessary to measure a level exceeding a certain number rather than an exact number of vermin when the government notifies alarms about them, such as jellyfish counting in the South Korea [19]. Also, in order to obtain an accurate outline in the FCN, it is necessary to have many training sets. It is possible to implement the desired function with a relatively small amount of data if the FCN is only trained enough to generate a good probability map instead of the outline proposal and combined with the regression network.
The processing time required for the inference was also decreased. For the processing of 80 images, the proposed network required 2.47 s on a platform with an Intel i7-6700K CPU and 2 Nvidia GTX 1080 GPUs. The conventional classification network took 4.64 s to process approximately 200 mosquitoes in the 80 images. Though the depth of the network is deeper with the FCN and neural network-based regression, two inferences require much less time when compared to multiple AlexNet inferences.
The proposed system is aimed at optimizing network usage for operation in embedded systems. Therefore, lighter network structures are preferred. The effort and human labor required to generate training data are also decreased by using the proposed pipeline as compared to the object detection network which requires the exact boxing of every mosquito in the image.

4. Conclusions and Future Work

This paper presents the development of an automated mosquito detection and control system based on the image processing of lured mosquitoes. The system can detect mosquitoes and release a Bti drug package into a target area, for example, sewage, if it is considered to be a mosquito larvae habitat. Thus, control of the mosquito population is more efficient than only removing mosquitoes at the imago stage. We expect that this system could be used not only in indoor urban areas, but also at rainfall collection facilities found in the tropical regions of developing countries.
In the repeated experiment, we could not find insects other than mosquitoes in our experimental site. However, for generality, we need to distinguish mosquitoes from any other kinds of insects which can be found outdoors. This will be a future task of our research.
As part of future research, we are considering optimizing the multiple network stacked structure. We are also considering benchmarking end-to-end structures and recent YOLO networks [20,21,22] to increase the object detection performance. Replacing AlexNet with ResNet to make the system faster is being considered. In addition, attaching the sensor node to a mobile system, such as a drone, is also being considered.

Author Contributions

K.K. and J.H. conducted research and revised the article; H.K. and H.L. aided the experiments under supervision of H.M.

Funding

This research was supported by Institute for Information and Communications Technology Promotion (IITP) grant funded by the Korean government (MSIT) (No. 2016-0-00563, Research on Adaptive Machine Learning Technology Development for Intelligent Autonomous Digital Companion).

Acknowledgments

This work was partially supported by BK21+.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Thomas, T.; De, T.D.; Shrma, P.; Lata, S.; Saraswat, P.; Pandey, K.C.; Dixit, R. Hemocytome: Deep sequencing analysis of mosquito blood cells in Indian malarial vector Anopheles stephensi. Gene 2016, 585, 177–190. [Google Scholar] [CrossRef] [PubMed]
  2. Pastula, D.M.; Smith, D.E.; Beckham, J.D.; Tyler, K.L. Four emerging arboviral diseases in North America: Jamestown Canyon, Powassan, chikungunya, and Zika virus diseases. J. Neurovirol. 2016, 22, 257–260. [Google Scholar] [CrossRef] [PubMed]
  3. Biswas, A.K.; Siddique, N.A.; Tian, B.B.; Wong, E.; Caillouët, K.A.; Motai, Y. Design of a Fiber-Optic Sensing Mosquito Trap. IEEE Sens. J. 2013, 13, 4423–4431. [Google Scholar] [CrossRef]
  4. Seo, M.J.; Gil, Y.J.; Kim, T.H.; Kim, H.J.; Youn, Y.N.; Yu, Y.M. Control Effects against Mosquitoes Larva of Bacillus thuringiensis subsp. israelensis CAB199 isolate according to Different Formulations. Korean J. Appl. Entomol. 2010, 49, 151–158. [Google Scholar] [CrossRef]
  5. Kim, Y.K.; Lee, J.B.; Kim, E.J.; Chung, A.H.; Oh, Y.H.; Yoo, Y.A.; Park, I.S.; Mok, E.K.; Kwon, J. Mosquito Prevalence and Spatial and Ecological Characteristics of Larval Occurrence in Eunpyeong Newtown, Seoul Korea. Kor. J. Nat. Conserv. 2015, 9, 59–68. [Google Scholar] [CrossRef]
  6. Burke, R.; Barrera, R.; Lewis, M.; Kluchinsky, T.; Claborn, D. Septic tanks as larval habitats for the mosquitoes Aedes aegypti and Culex quinquefasciatus in Playa-Playita, Puerto Rico. Med. Vet. Entomol. 2010, 24, 117–123. [Google Scholar] [CrossRef]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS’12), Lake Tahoe, ND, USA, 3–6 December 2012; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates Inc.: Red Hook, NY, USA, December 2012; pp. 1097–1105. [Google Scholar]
  8. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. Available online: https://arxiv.org/abs/1409.1556 (accessed on 23 August 2017).
  9. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  10. Kim, H.; Koo, J.; Kim, D.; Jung, S.; Shin, J.U.; Lee, S.; Myung, H. Image-Based Monitoring of Jellyfish using Deep Learning Architecture. IEEE Sens. J. 2016, 16, 2215–2216. [Google Scholar] [CrossRef]
  11. Motta, D.; Santos, A.Á.B.; Winkler, I.; Machado, B.A.S.; Pereira, D.A.D.I.; Cavalcanti, A.M.; Fonseca, E.O.L.; Kirchner, F.; Badaró, R. Application of convolutional neural networks for classification of adult mosquitoes in the field. PLoS ONE 2019, 14, e0210829. [Google Scholar] [CrossRef] [PubMed]
  12. Fuchida, M.; Pathmakumar, T.; Mohan, R.; Tan, N.; Nakamura, A. Vision-based perception and classification of mosquitoes using support vector machine. Appl. Sci. 2017, 7, 51. [Google Scholar] [CrossRef]
  13. Zhong, Y.; Gao, J.; Lei, Q.; Zhou, Y. A vision-based counting and recognition system for flying insects in intelligent agriculture. Sensors 2018, 18, 1489. [Google Scholar] [CrossRef] [PubMed]
  14. Hapairai, L.K.; Plichart, C.; Naseri, T.; Silva, U.; Tesimale, L.; Pemita, P.; Bossin, H.C.; Burkot, T.R.; Ritchie, S.A.; Graves, P.M.; et al. Evaluation of traps and lures for mosquito vectors and xenomonitoring of Wuchereria bancrofti infection in a high prevalence Samoan Village. Parasit. Vectors 2015, 287, 1–9. [Google Scholar] [CrossRef] [PubMed]
  15. Moore, S.J.; Du, Z.W.; Zhou, H.N.; Wang, X.Z.; Li, H.B.; Xiao, Y.J.; Hill, N. The efficacy of different mosquito trapping methods in a forest-fringe village, Yunnan Province, Southern China. Southeast Asian J. Trop. Med. Public Health 2001, 32, 282–289. [Google Scholar] [PubMed]
  16. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Available online: https://arxiv.org/abs/1409.0575 (accessed on 23 August 2017).
  17. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. Available online: https://arxiv.org/abs/1605.06211 (accessed on 23 August 2017).
  18. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding. Available online: https://arxiv.org/abs/1408.5093 (accessed on 23 August 2017).
  19. Kim, K.; Myung, H. Autoencoder-Combined Generative Adversarial Networks for Synthetic Image Data Generation and Detection of Jellyfish Swarm. IEEE Access 2018, 6, 54207–54214. [Google Scholar] [CrossRef]
  20. Redmon, J.; Divvala, S.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. Available online: https://arxiv.org/abs/1506.02640 (accessed on 23 August 2017).
  21. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. Available online: https://arxiv.org/abs/1612.08242 (accessed on 23 August 2017).
  22. Hoiem, D.; Chodpathumwan, Y.; Dai, Q. Diagnosing Error in Object Detectors. In Proceedings of the 12th European Conference on Computer Vision—Volume Part III (ECCV’12), Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 340–353. [Google Scholar] [CrossRef]
Figure 1. Overall procedure of the proposed mosquito detection and control system.
Figure 1. Overall procedure of the proposed mosquito detection and control system.
Sensors 19 02785 g001
Figure 2. (a) The mosquito control system developed. (b) Heat pad and Ultra Violet Light-Emitting Diode (UV LED) attached to the mosquito observation pad. (c) Close-up view of the automatic Bti larvicide dispenser.
Figure 2. (a) The mosquito control system developed. (b) Heat pad and Ultra Violet Light-Emitting Diode (UV LED) attached to the mosquito observation pad. (c) Close-up view of the automatic Bti larvicide dispenser.
Sensors 19 02785 g002
Figure 3. Flow diagram of the deep learning-based mosquito counting pipeline. The red arrows indicate data transfer and the blue arrows indicate network input/output. (a) Patch classifier network: mosquito colored-only image preparation using the pre-trained image patch classification network. (b) Score map network: FCN training based on the generated label from the previous stage. (c) Mosquito counting network: regression using a score map extracted from the trained FCN.
Figure 3. Flow diagram of the deep learning-based mosquito counting pipeline. The red arrows indicate data transfer and the blue arrows indicate network input/output. (a) Patch classifier network: mosquito colored-only image preparation using the pre-trained image patch classification network. (b) Score map network: FCN training based on the generated label from the previous stage. (c) Mosquito counting network: regression using a score map extracted from the trained FCN.
Sensors 19 02785 g003
Figure 4. (a) Mosquitoes lured by the sensor nodes. (b) Boxplot showing the number of mosquitoes observed in the recorded video. Ten random screenshots from 10 videos (trials) were used.
Figure 4. (a) Mosquitoes lured by the sensor nodes. (b) Boxplot showing the number of mosquitoes observed in the recorded video. Ten random screenshots from 10 videos (trials) were used.
Sensors 19 02785 g004
Figure 5. Input images (left) for the FCN and generated score maps (right) showing the probability of pixel matches to the presence of a mosquito. Each label indicates the number of mosquitoes in the image. (a) label 0, (b) label 1, (c) label 2, (d) label 3, (e) label 4, (f) label 5 (this label includes > 5 images as well).
Figure 5. Input images (left) for the FCN and generated score maps (right) showing the probability of pixel matches to the presence of a mosquito. Each label indicates the number of mosquitoes in the image. (a) label 0, (b) label 1, (c) label 2, (d) label 3, (e) label 4, (f) label 5 (this label includes > 5 images as well).
Sensors 19 02785 g005
Figure 6. Red circles indicate the ground truth value—the real number of mosquitoes in the image. The green line represents the estimation results from the FCN and neural network-based regression. The blue dotted line shows the number of mosquitoes detected by classifying every contour in the image using the classification network. The example of the misclassification image patch which causes the overestimate of the mosquito detection is represented on the left side.
Figure 6. Red circles indicate the ground truth value—the real number of mosquitoes in the image. The green line represents the estimation results from the FCN and neural network-based regression. The blue dotted line shows the number of mosquitoes detected by classifying every contour in the image using the classification network. The example of the misclassification image patch which causes the overestimate of the mosquito detection is represented on the left side.
Sensors 19 02785 g006
Table 1. A summary of the proposed network architecture based on the Fully Convolutional Network (FCN).
Table 1. A summary of the proposed network architecture based on the Fully Convolutional Network (FCN).
NameTypeInput SizeOutput SizeKernel SizeStride# of Filters
datadata3 × 500 × 5003 × 500 × 500
conv1convolution3 × 500 × 50064 × 500 × 5003 2
pool1max pooling64 × 500 × 50064 × 250 × 250221
conv2convolution128 × 250 × 250128 × 250 × 2503 2
pool2max pooling128 × 250 × 250128 × 125 × 125221
conv3convolution256 × 125 × 125256 × 125 × 1253 3
pool3max pooling256 × 125 × 125256 × 63 × 63221
conv4convolution512 × 63 × 63512 × 63 × 633 3
pool4max pooling512 × 63 × 63512 × 32 × 32221
conv5convolution512 × 32 × 32512 × 32 × 323 3
pool5max pooling512 × 32 × 32512 × 16 × 16221
fc6convolution512 × 16 × 164096 × 10 × 107 1
drop6dropout (rate 0.5)4096 × 10 × 104096 × 10 × 10
fc7convolution4096 × 10 × 104096 × 10 × 101 1
drop7dropout (rate 0.5)4096 × 10 × 104096 × 10 × 10
scoreconvolution4096 × 10 × 1021 × 10 × 101 1
score2deconvolution21 × 10 × 1021 × 22 × 22421
score-pool4convolution512 × 32 × 3221 × 32 × 321 1
score-pool4ccrop21 × 32 × 3221 × 22 × 22
score-fuseeltwise21 × 22 × 2221 × 22 × 22
bigscoredeconvolution21 × 22 × 2221 × 368 × 36832161
upscorecrop21 × 368 × 36821 × 500 × 500
outputsoftmax21 × 500 × 50021 × 500 × 500
Table 2. A summary of the proposed network architecture based on AlexNet (LRN: Local Response Normalization).
Table 2. A summary of the proposed network architecture based on AlexNet (LRN: Local Response Normalization).
NameTypeInput SizeOutput SizeKernel SizeStride# of Filters
datadata3 × 227 × 2273 × 227 × 227
conv1convolution3 × 227 × 22796 × 55 × 551141
norm1LRN96 × 55 × 5596 × 55 × 55
pool1max pooling96 × 55 × 5596 × 27 × 27321
conv2convolution96 × 27 × 27256 × 27 × 275 1
norm2LRN256 × 27 × 27256 × 27 × 27
pool2max pooling256 × 27 × 27256 × 13 × 13321
conv3convolution256 × 13 × 13384 × 13 × 133 1
conv4convolution384 × 13 × 13384 × 13 × 133 1
conv5convolution384 × 13 × 13256 × 13 × 133 1
pool5max pooling256 × 13 × 13256 × 6 × 6321
fc6InnerProduct256 × 6 × 64096 × 1 × 1 1
drop6dropout (rate 0.5)4096 × 1 × 14096 × 1 × 1
fc7InnerProduct4096 × 1 × 14096 × 1 × 1 1
drop7dropout (rate 0.5)4096 × 1 × 14096 × 1 × 1
fc8InnerProduct4096 × 1 × 11000 × 1 × 1 1
lossSoftmaxWithLoss1000 × 1 × 11000 × 1 × 1

Share and Cite

MDPI and ACS Style

Kim, K.; Hyun, J.; Kim, H.; Lim, H.; Myung, H. A Deep Learning-Based Automatic Mosquito Sensing and Control System for Urban Mosquito Habitats. Sensors 2019, 19, 2785. https://doi.org/10.3390/s19122785

AMA Style

Kim K, Hyun J, Kim H, Lim H, Myung H. A Deep Learning-Based Automatic Mosquito Sensing and Control System for Urban Mosquito Habitats. Sensors. 2019; 19(12):2785. https://doi.org/10.3390/s19122785

Chicago/Turabian Style

Kim, Kyukwang, Jieum Hyun, Hyeongkeun Kim, Hwijoon Lim, and Hyun Myung. 2019. "A Deep Learning-Based Automatic Mosquito Sensing and Control System for Urban Mosquito Habitats" Sensors 19, no. 12: 2785. https://doi.org/10.3390/s19122785

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop