Next Article in Journal
A Bidirectional Mode-Locked Fiber Laser for Asynchronous Noise-like Pulse Generation
Previous Article in Journal
Low Complexity Parallel Carrier Frequency Offset Estimation Based on Time-Tagged QPSK Partitioning for Coherent Free-Space Optical Communication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Long-Range Imaging through Scattering Media Using Deep Learning

1
College of Engineering Physics, Shenzhen University of Technology, Shenzhen 518118, China
2
College of Applied Technology, Shenzhen University, Shenzhen 518060, China
*
Author to whom correspondence should be addressed.
Photonics 2024, 11(9), 887; https://doi.org/10.3390/photonics11090887
Submission received: 16 August 2024 / Revised: 4 September 2024 / Accepted: 10 September 2024 / Published: 20 September 2024

Abstract

:
Imaging through scattering media is an important and challenging problem, and the technology has been used in many fields, such as autonomous driving, industrial inspections, remote sensing imaging, and biomedical imaging. However, most of the previous experiments used numbers or letters for close-range imaging, while objects in life are colorful. In this study, a new deep learning network, DesUNet, was constructed to image realistic objects at medium and long distances under sunlight through scattering media, and to realize object recognition. In addition, this study also compares the imaging results of different neural networks, and the results show that the DesUNet network improves the feature information storage ability and enhances the image reconstruction. It not only clearly restores the original appearance of the object, but also extracts the physical information about the object. In order to further verify the power of the DesUNet network, this study also conducted indoor near distance and outdoor medium distance imaging experiments. For indoor reconstructed objects, the appearance of the objects could be clearly identified. For outdoor reconstructed objects, the confidence level could reach above 0.9 through YOLO. The experiments show that the DesUNet network has good robustness and generalization.

1. Introduction

Light is a powerful means of imprinting and recording the characteristics of objects in real time on rewritable molds [1]. The different properties of light, such as intensity and phase distributions, polarization and spectra, allow us to perceive the reflectance and thickness distributions, as well as the birefringence and spectral absorption properties of an object [2]. When light interacts with objects, it is usually characterized by multiple scattering and absorption events [3]. When light passes through some fog, heavy rain, clouds, biological tissues, and other strong scattering media, inter-particle interactions change the wavefront phase of the incident light. As a result, the original orderly light field of the incident light becomes chaotic, which ultimately leads to a loss of information about the object.
The rapid development of photonics has accelerated the development of scattering optics techniques in optical imaging [4]. The simplest and most common form of optical imaging is the use of lenses to reproduce the intensity of light scattered by an object on an imaging plane, where it can be measured by a high-pixel detector [5]. For the inverse problem of scattering medium imaging, if light is scattered during propagation, the wavefront phase of the incident light changes due to inter-particle interactions, resulting in the originally ordered light field of the outgoing light becoming chaotic, and the light from each point on the object will be dispersed to multiple detector pixels. This leads to optical model errors and uncertainties in the imaging process, as well as the need to process a large amount of scattering data, thus requiring the support of high-performance computing devices. In addition, practical applications may be interfered by ambient light, requiring appropriate optical design and processing methods to reduce the interference. These limitations make it difficult for conventional imaging methods to effectively solve the inverse problem of scattering medium imaging.
In order to solve these limitations, researchers have proposed a variety of scattered light imaging methods, such as optical phase conjugate scattering imaging technology [6] and optical transmission matrix scattering imaging technology [7,8]. Based on the optical memory effect scattering imaging technology [9,10,11], researchers have used these techniques to achieve focusing or imaging through the scattering medium. However, each of these methods has its own advantages, disadvantages, and applicable conditions, such as the optical memory effect based scattering imaging technique failing in the face of thick scattering media, and also failing when the size of the observed object is large. In particular, when scattered by thick scatterers, the signal of ballistic light is very weak and the scattering ‘noise’ very strong, which makes the classical optimization algorithms for solving inverse problems unable to help.
Recent studies have shown that when deep learning algorithms are used instead of the traditional inverse problem [12,13,14], they are able to effectively recover object images from recorded scattering, thus overcoming the limitations of traditional methods and are faster, more reliable, and more accurate for data processing, image segmentation, and image reconstruction [15,16,17,18,19,20,21,22,23,24]. They are powerful in solving complex computational imaging problems [25], super-resolution imaging [26], and polarization imaging [27]. With the rapid development of artificial intelligence, deep learning algorithms are making huge improvements to optical imaging [3,28,29,30,31,32,33,34]. In terms of imaging, the earliest study by Lyu et al. completed the use of deep learning through scattering medium imaging [35]. Although the construction of the neural network is very simple, and can only carry out a simple digital reconstruction, this opens up a new path of scattering imaging technology. Subsequently, Li proposed the use of convolutional neural networks to achieve coherent imaging through thin frosted glass [36], and one year later, Lyu also proposed a hybrid neural network for imaging experiments through scattering media [37], which further enriched the through scattering media imaging technology by improving on the original neural network. Ma et al. proposed a plug-and-play algorithm based on a generalized alternating projection optimization framework, and in order to be able to achieve imaging through scattering media in an interfering environment, they combined this algorithm with neural networks and the Fienup phase recovery method [38]. Cheng et al. took the scattered autocorrelation information as the physical constraint, proposed a two-level neural network that realizes background light denoising and object reconstruction through deep learning, and constructed an end-to-end deep neural network, thus overcoming the interference of background light, and the interference of object reconstruction [39]. Zheng et al. built an end-to-end deep neural network that can overcome the interference of background light [40,41], while Zhang et al. designed a space-frequency dual-domain deep neural network using the frequency-domain characteristics of speckle [42]. However, current deep learning-based approaches are still poor in their ability to generalize to unknown scenarios. In 2021, the Zheng research group further realized underwater dynamic incoherent scattering imaging [40], in which the researchers put a fat milk solution in a water tank to simulate the dynamic scattering scenario. In the experiment, in addition to testing the traditional MINST digital set, the image recovery of more complex gray objects was also realized, and a two-camera method was proposed to measure the dynamic characteristics of scattering media (decoherence time). In 2023, Liu et al. proposed a method to obtain a quantitative scattering fog pattern based on a passive luminescence electronic ink screen to solve the problem of data acquisition [43]. The E ink screen simulates passively glowing natural objects to obtain data pairs for training the network and testing fog maps in real scenes. However, the above neural networks all had defects:, the feature storage ability of the image was relatively weak, their experimental objects were simple numbers and letters, for complex real objects, and the neural network they built may not be able to achieve a clear reconstruction. Because real objects are complex and changeable, with more feature information, it is necessary for neural networks to store a large amount of feature information, numbers and letters have simple structures and obvious features, and the performance of neural networks is not high. The traditional neural network has many problems. Even if the training can eventually converge, the existence of scattering medium leads to excessive noise interference in the image, which may lead to an overfitting phenomenon [44]. Due to the weak ability of storing object feature information, the trained network is difficult to generalize and the final imaging effect is poor.
The neural network used in this study will focus on solving these problems, using deep learning to improve the U-net neural network, which is called DesUNet here. In this study, a real scene is taken as the target object for experiments, and the scattering medium imaging is really applied to life. Through the experimental results, it can be proved that the DesUNet neural network has better performance for storing feature information. Firstly, indoor objects are used for imaging through scattering media, and the reconstruction results of other traditional neural networks are constructed and compared with the DesUNet neural networks, which proves that the DesUNet neural network is better than other neural networks in storing feature information. After that, the application scene will be transferred to the outdoors. In contrast to other imaging experiments through scattering media, the target objects used in this study were people, cars and other complex objects in real scenes. This neural network is used to image through the scattering medium in the middle and long distance (1–2 km), complete the experiment, and recognize the reconstructed object. Through indoor and outdoor experiments, it was finally verified that the neural network had better performance for storing feature information.

2. Rationale

2.1. Construction of the DesUNet Neural Network

Scattering imaging is a nonlinear reversible problem that is addressed using deep learning, which compensates for the shortcomings of traditional imaging methods such as complexity and high computational demands. By harnessing the computational power and fitting capability of deep learning, it facilitates the fitting calculation of scattering media and the final imaging results. The mathematical model of its imaging system is as follows:
OP = F(IP)
where OP is the output of the scattering imaging system, IP is the input of the scattering imaging system, and F is a positive function of the input light field to the output light field.
Since the optical path is reversible and the absorption of light by the scattering medium is neglected, the mathematical model from the output light field to the input light field is as follows:
I′P = F−1(O′P)
where I′P is the input to the inverse process, O′P is the output of the inverse process, and F−1 is the inverse of F.
The result is the ideal mathematical model for the problem of imaging through scattering media. As shown in Equations (1) and (2) above, there is a correspondence between the scattering image and the original image of the target object, and the functions are handed over to deep learning, since the ultimate goal of imaging through scattering media is to make the inverse input approximately equal to the positive input.
Due to the weak storage ability of traditional neural networks for object feature information and to solve this problem, a new neural network named “DesUNet” was builtin this study. The structure of the constructed DesUNet neural network is similar to that of the U-Net neural network [45], as shown in Figure 1. A U-shaped architecture was made to connect low-level features with high-level features, enabling better capture of features at different levels. Due to the network’s architecture, it can be divided into two parts: the down-sampling process and the up-sampling process. The down-sampling process reduces the dimensions of the image and retains valuable information to some extent, helping to prevent overfitting. The up-sampling process then restores the reduced image to its original size. The purpose of building the DesUNet neural network is to increase the storage capacity of the object feature information. Considering this, a dense block was built, as shown in Figure 2. Its role is to stack the features extracted by each layer of the neural network that precedes it. Feature maps can be reused between different layers of the network, enhancing the flow of gradients, thus improving feature utilization efficiency and greatly enhancing the model’s performance and generalization capabilities. Importantly, the feature information from each layer is fully utilized. For example, the feature information from the first layer is fused with that from the second layer, the feature information from the first layer is fused with that from the third layer, and the feature information fused in the previous two steps is fused again. Each dense block consists of Batch Normalization layers, Activation layers (using ReLU as the activation function), Convolutional layers (Conv layers), Dropout layers (used to reduce overfitting), and Concatenate layers. The aim is to obtain more target object feature information. The input is the original object image with dimensions of 256 × 256 pixels, along with the corresponding speckle image. These inputs go through four Conv layers, dense blocks, and Max Pooling layers, resulting in a reduction in the size of the target object and feature extraction. During image processing in the neural network, the smallest image size is 16 × 16 pixels, but it has a high channel count of 1088, which contains rich feature information. This is 10–20 times more feature information than conventional neural networks, making it much better suited to handling complex objects. The inclusion of Batch Normalization at each layer significantly improves the training speed of the network, and the network’s generalization capability is further strengthened. Following this, a path consisting of four Conv layers, dense blocks, and Concatenate layers (including deconvolution operations to restore the image to its original size and improve the neural network’s accuracy and efficiency) is utilized. Subsequently, through Conv layers, dense blocks, and Conv layers, the output goes to the final layer. The final layer’s convolution integrates the results of processing the input data across all previous layers, ultimately generating the reconstructed result image.
Compared with other neural networks, the DesUNet neural network has several advantages. First, due to its structural design, it can combine low-level feature information with high-level feature information, and this ability enables it to better represent and understand the low-level features of complex objects, which makes it better able to capture the features of target objects. Secondly, due to the addition of the dense block, the neural network can improve the efficiency of feature utilization and retain the spatial information of the image. Third, it can provide more powerful feature expression capabilities, resulting in more accurate reconstruction results and image segmentation. Fourth, it can reduce the computational complexity of the network while maintaining high performance, and improve the computational efficiency of the network, which is very valuable for practical applications requiring real-time or resource-efficient processing.

2.2. Experimental Setup

The experimental setup, as shown in Figure 3, differs from most imaging through scattering media experiments, which typically involve illuminating the target object with lasers to enhance its features. In this study, experiments were conducted under natural light (sunlight) conditions to better simulate real-world scenarios and demonstrate the powerful feature processing capabilities of the DesUNet neural network. The detector used is an integrated photoelectric sensor mounted on a platform (HP-DMA2186), as depicted in Figure 4. The scattering medium in the imaging system comprise 2 mm ground glass diffusers (polishing specifications: 220 grits), which are pressed against the lens of the detector. The target object under investigation can be any arbitrary item. The indoor object is placed about 5–10 m away from the detector, and the outdoor object is 1–2 km away from the detector. This experimental configuration allows for imaging through scattering media using natural light, aiming to showcase the DesUNet neural network’s ability to handle feature-rich information in a real-world setting.

2.3. Indoor Data Acquisition and Processing

Firstly, the imaging through scattering medium experiments were conducted indoors in order to fully prepare for the outdoor experiments, and the most targeted object was a schoolbag for testing (the schoolbag was chosen because of its distinctive features), and data collection was conducted for more than 300 h. In order to collect the training data, the object was placed at different distances and different postures, and three groups of experiments were conducted, where the bag was placed 5 m, 7 m, and 10 m away from the detector for data collection. In each group, 8000 original images of the object and 8000 corresponding scattered images were collected and the data used as training data. All images were 256 × 256 pixels and all were in color. The choice of color images aimed to simulate real-world scenarios more accurately, as color images contain richer information. It is worth noting that the “cv2.imread” function in Python reads images in the BGR format (where B stands for blue, G for green, and R for red), so the displayed image colors may appear differently. To address this, the final reconstructed results were adjusted to be displayed in a heatmap-like form and in grayscale.

3. Experimental Results and Analysis

3.1. Scattering Imaging of Indoor Objects

The reconstructed results obtained from training the test dataset with the DesUNet neural network, as shown in Figure 4, appear quite satisfactory. Comparing the original images with the reconstructed ones, it is evident that DesUNet excels in handling details, and is particularly successful in reconstructing the edges of the objects. From the indoor experiments, it can be concluded that DesUNet demonstrates excellent object reconstruction capabilities. However, it is important to note that this alone does not imply that other neural networks cannot perform well in this experiment. It showcases DesUNet’s strong reconstruction ability but doesn’t directly prove its superior feature information storage capability.
To further demonstrate the high performance of the DesUNet neural network, a straightforward approach is to conduct comparative experiments. Specifically, traditional CNN neural networks and U-Net neural networks were set up, and all three neural networks were used for imaging through scattering media experiments. The reconstructed results from these experiments are shown in Figure 5. From the results, it is evident that imaging through scattering media yields different outcomes for the same object at different distances and with different neural networks. Among them, the DesUNet network exhibited the best imaging performance. The reconstructed results based on the traditional CNN neural network, while recognizing the general shape of the object, appeared highly blurred and only showed the rough outline. This indicates the weaker feature processing capability of traditional CNN neural networks. On the other hand, the reconstructed results based on the U-Net neural network, while revealing the basic appearance of the object, lacked clear edge information. This suggests that although the U-Net neural network can handle some feature information, it falls short of the expected results for this experiment. It is important to note that this is a simple backpack, and outdoor scenes are more complex, so achieving such results may not always be possible. In contrast, the reconstructed results based on the DesUNet neural network displayed sharp edge contours and clear object features, essentially restoring the original appearance of the object. Through indoor experiments, it can be concluded that both the U-net neural network and DesUNet neural network have certain capabilities of processing feature information, and to prove that the DesUNet neural network has a stronger capability of processing feature information, it is necessary to conduct imaging experiments through scattering media in outdoor real scenes.
The comparison between the original image of the target object and the reconstruction results of different neural networks was carried out by utilizing SSIM (structural similarity), assuming that the two input images are X and Y, where the formula for SSIM is as follows:
SSIM X , Y = 2 μ X μ Y + C 1 2 δ X Y + C 2 μ X 2 μ Y 2 + C 1 δ X 2 δ Y 2 + C 2
where μX and μY denote the mean of X and Y respectively, δX2 and δY2 denote the variance of X and Y respectively, δXY denotes the covariance of X and Y, C1 and C2 in order to prevent the denominator from being a constant of 0. The value of SSIM is in the range of [0, 1], and the larger the value means the smaller the gap between the two plots. The final comparison result graph is shown in Figure 6. According to the SSIM score, the reconstruction result based on the DesUNet neural network is the closest to the original image.

3.2. Scattered Imaging Outdoors

By comparing indoor experiments, it can be concluded that the DesUNet neural network exhibits good performance. However, indoor experiments involving simple target objects do not necessarily indicate that DesUNet will perform similarly in real-world scenarios. To address this, outdoor experiments were conducted. Real-world applications often involve complex and diverse scenes, far from just simple digits and letters. To better simulate real scenarios, this experiment used real outdoor backgrounds with vehicles and pedestrians as target objects for imaging through scattering media. In the outdoor environment, we collected a large number of outdoor road pictures and scattered images as a training set. The training data content included three major categories of people, cars and the road environment, where each person’s appearance and morphology were different, and cars were divided into cars, electric cars, bicycles and other car vehicles. In order for the experiment to be closer to the real situation, the data collection time spanned more than 1000 h, because every day’s road conditions may be different. eEach set of photos were color RGB images, through a long period of training and we finally selected one of them as the training results, as shown in Figure 7. It is evident that the real road conditions’ image and the reconstructed image bear a striking resemblance, illustrating the DesUNet neural network’s ability to perform imaging through scattering media in realistic outdoor scenarios.
The training results met expectations, and subsequently, speckle images of different targets were collected as a test dataset for imaging through scattering media experiments. The targets included people, cars, motorcycles, and bicycles. The reconstructed results for different neural networks, focusing on the target regions, are presented in Figure 8. Notably, the target objects are located at distances of 1–2 km from the detector. It is evident that the reconstructed results based on the DesUNet neural network were quite satisfactory. They provided a clear view of the target objects, demonstrating that the DesUNet neural network had the capability to perform imaging through scattering media in complex outdoor environments and at medium to long distances. For the sake of experimental completeness, imaging through scattering media experiments were also conducted with CNN and U-Net neural networks. However, it is clear that these two neural networks, while able to reconstruct the backpack indoors, struggled to perform well when the target objects were moved outdoors. In outdoor scenarios, these two neural networks fell short of the expected results, emphasizing the DesUNet neural network’s superiority in such challenging conditions.

3.3. Recognize Objects after Imaging through a Diffuse Image

To test the accuracy of the neural network in this study, the YOLO (You Only Look Once) algorithm was used for object recognition in the reconstructed results. The YOLO algorithm is based on a single neural network and performs region proposal generation. It does not differentiate between foreground and background but instead classifies and regresses the target objects. One major advantage of YOLO is its ability to significantly improve object detection speed while reducing the number of object parameters and computational load. Using YOLO for object recognition in the reconstructed results based on the DesUNet neural network, the recognition results are shown in Figure 9. Observing the confidence of target objects, the confidence includes two aspects. One aspect is the likelihood that the bounding box contains an object (Pr(object)), where Pr(object) = 1 when the object is present, and Pr(object) = 0 when the object is absent. The other aspect is the accuracy of the bounding box, represented by the Intersection over Union (IoU) between the predicted box and the actual box. The confidence is displayed after the predicted object name for easy observation. From the results, it is clear that the target objects in the reconstructed images were accurately identified.
Through the experiments conducted, this study successfully demonstrated the ability to perform imaging through scattering media in complex outdoor environments and at medium to long distances. It also highlighted the high performance of the DesUNet neural network. In combination with this study, it becomes evident that imaging through scattering media technology can be applicable in real-world environments, extending beyond simple digit and letter reconstruction. This has significant implications for various fields, including transportation, outdoor scenarios, and military applications, where imaging through scattering media can play a crucial role.

4. Conclusions

In the context of imaging through scattering media technology, the experiments conducted in this study have demonstrated that the DesUNet neural network significantly improves the accuracy of reconstructed images through scattering media. The incorporation of dense blocks in the DesUNet network serves multiple purposes, including preventing gradient explosion or vanishing, enhancing network generalization capabilities, and ultimately resulting in clearer and more distinct reconstructed images through scattering media. Comparative experiments were conducted with traditional CNN and U-Net neural networks, and the results clearly show that traditional neural networks produce highly blurred reconstructions, especially in complex conditions. In contrast, the DesUNet neural network not only achieves clear reconstructions at short distances but also excels in reconstructing images through scattering media under complex outdoor conditions and at medium to long distances, coupled with object recognition. The DesUNet neural network has successfully demonstrated imaging through scattering media in real-world environments. This technology has versatile applications across various fields. In the transportation sector, for instance, in vehicular or stationary settings, it can provide clear visibility of the road ahead, even in foggy conditions, allowing for immediate adjustments, and reducing the occurrence of traffic accidents. In outdoor scenarios with complex environments (such as forests or cloudy skies), where much of the field of view is obstructed, this technology can be employed for precise localization and object identification. In future research work, we will continue to optimize the neural network of reconstructed objects after scattering and further improve the network structure. Of course, in addition to optimizing the deep learning algorithm, the design of optical imaging systems is also very important. It is important to obtain more dimensions of optical field data or realize pre-processing methods of optical field regulation to break through the lower limit of imaging depth. On this basis, a more advanced algorithm is designed to post-process the acquired multidimensional data, which can further improve the recovery effect. With the development of optoelectronic devices and the improvement of algorithm computing power, scattering imaging research is expected to further organically combine the front and back ends, stimulate new ideas, deal with more complex application scenarios and objects, and achieve further improvements in imaging depth and recovery quality.

Author Contributions

Conceptualization, C.Z. and W.D.; methodology, W.D. and Y.J.; software, Y.J.; validation, W.D. and Y.J.; formal analysis, W.D. and Y.J.; investigation, W.D. and Y.J.; resources, C.Z. and W.D.; data curation, W.D. and Y.J.; writing—original draft preparation, Y.J.; writing review and editing, W.D. and Y.J.; visualization, Y.J.; supervision, W.D. and Y.J.; project administration, C.Z. and W.D.; funding acquisition, C.Z. and W.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shenzhen Science and Technology Program (Grant No. ZDSYS20200811143600001).

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rosen, J.; Alford, S.; Allan, B.; Anand, V.; Arnon, S.; Arockiaraj, F.G.; Art, J.; Bai, B.; Balasubramaniam, G.M.; Birnbaum, T.; et al. Roadmap on computational methods in optical imaging and holography [invited]. Appl. Phys. B 2024, 130, 166. [Google Scholar] [CrossRef] [PubMed]
  2. Boas, D.A.; Dale, A.M.; Franceschini, M.A. Diffuse optical imaging of brain activation: Approaches to optimizing image sensitivity, resolution, and accuracy. Neuroimage 2004, 23, S275–S288. [Google Scholar] [CrossRef] [PubMed]
  3. Balasubramaniam, G.M.; Biton, N.; Arnon, S. Imaging through diffuse media using multi-mode vortex beams and deep learning. Sci. Rep. 2022, 12, 1561. [Google Scholar]
  4. Zhang, X. Instrumentation in Diffuse Optical Imaging. Photonics 2014, 1, 9–32. [Google Scholar] [CrossRef]
  5. Bertolotti, J.; Katz, O. Imaging in complex media. Nat. Phys. 2022, 18, 1008–1017. [Google Scholar] [CrossRef]
  6. Yaqoob, Z.; Psaltis, D.; Feld, M.S.; Yang, C.H. Optical phase conjugation for turbidity suppression in biological samples. Nat. Photonics 2008, 2, 110–115. [Google Scholar] [CrossRef]
  7. Popoff, S.M.; Lerosey, G.; Carminati, R.; Fink, M.; Boccara, A.C.; Gigan, S. Measuring the transmission matrix in optics: An approach to the study and control of light propagation in disordered media. Phys. Rev. Lett. 2010, 104. [Google Scholar] [CrossRef]
  8. Popoff, S.M.; Lerosey, G.; Fink, M.; Boccara, A.C.; Gigan, S. Controlling light through optical disordered media: Transmission matrix approach. New J. Phys. 2011, 13, 123021. [Google Scholar] [CrossRef]
  9. Katz, O.; Heidmann, P.; Fink, M.; Gigan, S. Non-invasive single-shot imaging through scattering layers and around corners via speckle correlations. Nat. Photonics 2014, 8, 784–790. [Google Scholar] [CrossRef]
  10. Judkewitz, B.; Horstmeyer, R.; Vellekoop, I.M.; Papadopoulos, I.N.; Yang, C.H. Translation correlations in anisotropically scattering media. Nat. Phys. 2015, 11, 684–689. [Google Scholar] [CrossRef]
  11. Bertolotti, J.; Putten, E.G.; Blum, C.; Lagendijk, A.; Vos, W.L.; Mosk, A.P. Non-invasive imaging through opaque scattering layers. Nature 2012, 491, 232–234. [Google Scholar] [CrossRef] [PubMed]
  12. Suárez, P.L.; Sappa, A.D.; Vintimilla, B.X.; Hammoud, R.I. Deep learning based single image dehazing. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1250–12507. [Google Scholar]
  13. Gui, J.; Cong, X.; Cao, Y.; Ren, W.; Zhang, J.; Zhang, J.; Cao, J.; Tao, D. A comprehensive survey and taxonomy on single image dehazing based on deep learning. ACM Comput. Surv. 2023, 55, 279. [Google Scholar] [CrossRef]
  14. Jia, T.Y.; Zhuo, L.; Li, J.F.; Zhang, J. Research advances on deep learning based single image dehazing. Acta Electonica Sin. 2023, 51, 231–245. [Google Scholar]
  15. Fu, Z.F.; Wang, F.; Tang, Z.W.; Bian, Y.; Situ, G. Adaptive imaging through dense dynamic scattering media using transfer learning. Opt. Express 2024, 32, 13688–13700. [Google Scholar] [CrossRef]
  16. Yoo, J.; Sabir, S.; Heo, D.; Kim, K.H.; Wahab, A.; Choi, Y.; Lee, S.I.; Chae, E.Y.; Kim, H.H.; Bae, Y.M.; et al. Deep Learning Diffuse Optical Tomography. IEEE Trans. Med. Imaging 2020, 39, 877–887. [Google Scholar] [CrossRef]
  17. Survarachakan, S.; Pelanis, E.; Khan, Z.A.; Kumar, R.P.; Edwin, B.; Lindseth, F. Effects of enhancement on deep learning based hepatic vessel segmentation. Electronics 2021, 10, 1165. [Google Scholar] [CrossRef]
  18. Balasubramaniam, G.M.; Wiesel, B.; Biton, N.; Kumar, R.; Kupferman, J.; Arnon, S. Tutorial on the Use of Deep Learning in Diffuse Optical Tomography. Electronics 2022, 11, 305. [Google Scholar] [CrossRef]
  19. Ben Yedder, H.; BenTaieb, A.; Shokoufi, M.; Zahiremami, A.; Golnaraghi, F.; Hamarneh, G. Deep learning based image reconstruction for diffuse optical tomography. In Machine Learning for Medical Image Reconstruction; Springer: Cham, The Netherlands, 2018; Volume 11074, pp. 112–119. [Google Scholar]
  20. Applegate, M.B.; Istfan, R.E.; Spink, S.; Tank, A.; Roblyer, D. Recent advances in high speed diffuse optical imaging in biomedicine. APL Photonics 2020, 5, 040802. [Google Scholar] [CrossRef]
  21. Fang, X.; Gao, C.; Li, Y.; Li, T. Solving heterogenous region for diffuse optical tomography with a convolutional forward calculation model and the inverse neural network. In Proceedings of the Advanced Optical Imaging Technologies III—SPIE/COS Photonics Asia, Online, 11–16 October 2020; Volume 18. [Google Scholar]
  22. Kabe, G.K.; Song, Y.; Liu, Z. Optimization of fire net for liver lesion classification. Electronics 2020, 9, 1237. [Google Scholar] [CrossRef]
  23. Chen, Z.; Ma, G.; Jiang, Y.; Wang, B.; Soleimani, M. Application of Deep Neural Network to the Reconstruction of Two-Phase Material Imaging by Capacitively Coupled Electrical Resistance Tomography. Electronics 2021, 10, 1058. [Google Scholar] [CrossRef]
  24. Cho, C.; Lee, Y.H.; Park, J.; Lee, S. A self-spatial adaptive weighting based u-net for image segmentation. Electronics 2021, 10, 348. [Google Scholar] [CrossRef]
  25. Lin, H.; Huang, C.; He, Z.; Zeng, J.; Chen, F.; Yu, C.; Li, Y.; Zhang, Y.; Chen, H.; Pu, J. Phase Imaging through Scattering Media Using Incoherent Light Source. Photonics 2023, 10, 792. [Google Scholar] [CrossRef]
  26. Li, W.; Abrashitova, K.; Osnabrugge, G.; Amitonova, L.V. Generative Adversarial Network for Superresolution Imaging through a Fiber. Phys. Rev. Appl. 2022, 18, 034075. [Google Scholar] [CrossRef]
  27. Lin, B.; Fan, X.; Li, D.; Guo, Z. High-Performance Polarization Imaging Reconstruction in Scattering System under Natural Light Conditions with an Improved U-Net. Photonics 2023, 10, 204. [Google Scholar] [CrossRef]
  28. Sengupta, S.; Mittal, N.; Modi, M. Improved skin lesions detection using color space and artificial intelligence techniques. J. Dermatolog. Treat 2020, 31, 511–518. [Google Scholar] [CrossRef]
  29. Hattab, G.; Arnold, M.; Strenger, L.; Allan, M.; Arsentjeva, D.; Gold, O.; Simpfendörfer, T.; Maier-Hein, L.; Speidel, S. Kidney edge detection in laparoscopic image data for computer-assisted surgery: Kidney edge detection. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 379–387. [Google Scholar] [CrossRef]
  30. Baek, J.W.; Chung, K. Pothole classification model using edge detection in road image. Appl. Sci. 2020, 10, 6662. [Google Scholar] [CrossRef]
  31. Wetzstein, G.; Ozcan, A.; Gigan, S.; Fan, S.; Englund, D.; Soljačić, M.; Denz, C.; Miller, D.A.B.; Psaltis, D. Inference in artificial intelligence with deep optics and photonics. Nature 2020, 588, 39–47. [Google Scholar] [CrossRef]
  32. Barbastathis, G.; Ozcan, A.; Situ, G. On the use of deep learning for computational imaging. Optica 2019, 6, 921. [Google Scholar] [CrossRef]
  33. Nishijima, T.; Arimoto, H.; Watanabe, W. Image reconstruction behind diffuser by deep learning and spatial filtering. In Proceedings of the Biomedical Imaging and Sensing Conference 2020, Yokohama, Japan, 22–24 April 2020. [Google Scholar]
  34. Sharifi, S.; Banadaki, Y.; Veronis, G.; Dowling, J.P. Towards classification of experimental Laguerre-Gaussian modes using convolutional neural networks. Opt. Eng. 2020, 59, 076113. [Google Scholar] [CrossRef]
  35. Lyu, M.; Wang, H.; Li, G.; Situ, G. Exploit imaging through opaque wall via deep learning. arXiv 2017, arXiv:1708.07881. [Google Scholar]
  36. Li, S.; Deng, M.; Lee, J.; Sinha, A.; Barbastathis, G. Imaging through glass diffusers using densely connected convolutional networks. Optica 2018, 5, 803–813. [Google Scholar] [CrossRef]
  37. Lyu, M.; Wang, H.; Li, G.W.; Zheng, S.; Situ, G. Learning-based lensless imaging through optically thick scattering media. Adv. Photonics 2019, 1, 036002. [Google Scholar] [CrossRef]
  38. Ma, K.; Wang, X.; He, S.; Li, L. Plug-and-Play Algorithm for Imaging through Scattering Media under Ambient Light Interference. Opt. Lett. 2023, 48, 1754–1757. [Google Scholar] [CrossRef]
  39. Cheng, Q.; Guo, E.; Gu, J.; Bai, L.; Han, J.; Zheng, D. De-Noising Imaging through Diffusers with Autocorrelation. Appl. Opt. 2021, 60, 7686–7695. [Google Scholar] [CrossRef]
  40. Zheng, S.; Wang, H.; Dong, S.; Wang, F.; Situ, G. Incoherent imaging through highly non static and optically thick turbid media based on neural network. Photonics Res. 2021, 9, B220–B228. [Google Scholar] [CrossRef]
  41. He, S.; Wang, X.; Li, L. Polarized-Speckle Deviation Imaging through Scattering Media under Strong Background Light Interference. Photonics 2024, 11, 682. [Google Scholar] [CrossRef]
  42. Zhang, Y.; Huang, H.; Wu, F.; Han, J.; Yang, Y.; Li, R. Imaging through a Scattering Medium under Different Intensities of Ambient Light Interference. Photonics 2023, 10, 1023. [Google Scholar] [CrossRef]
  43. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. AOD-net: All-in-one dehazing network. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4780–4788. [Google Scholar]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  45. Weng, W.H.; Zhu, X. Inet: Convolutional networks for biomedical image segmentation. IEEE Access 2021, 9, 16591–16603. [Google Scholar] [CrossRef]
Figure 1. DesUNet neural network structure.
Figure 1. DesUNet neural network structure.
Photonics 11 00887 g001
Figure 2. Dense block structure.
Figure 2. Dense block structure.
Photonics 11 00887 g002
Figure 3. Experimental setup.
Figure 3. Experimental setup.
Photonics 11 00887 g003
Figure 4. Image result of different distance through scattering medium. (a) A speckle image; (b) The original drawing of the target object; (c) Restored image for the target object in the scattering medium.
Figure 4. Image result of different distance through scattering medium. (a) A speckle image; (b) The original drawing of the target object; (c) Restored image for the target object in the scattering medium.
Photonics 11 00887 g004
Figure 5. Comparison of imaging results of different locations through scattering media by different neural networks. (a,d) Speckle images; (b,e) The original drawing of the target object; (c,f) Reconstructed image for the target object in the scattering medium.
Figure 5. Comparison of imaging results of different locations through scattering media by different neural networks. (a,d) Speckle images; (b,e) The original drawing of the target object; (c,f) Reconstructed image for the target object in the scattering medium.
Photonics 11 00887 g005
Figure 6. SSIM was used to compare the original image with the reconstructed image.
Figure 6. SSIM was used to compare the original image with the reconstructed image.
Photonics 11 00887 g006
Figure 7. Road chart. (a) Environmental artwork; (b) Speckle images; (c) Reconstructed images.
Figure 7. Road chart. (a) Environmental artwork; (b) Speckle images; (c) Reconstructed images.
Photonics 11 00887 g007
Figure 8. The reconstruction results of the middle and long distance. (a) Speckle image corresponding to the target; (b) Reconstruction results based on DesUNet neural network; (c) Reconstruction results based on U-net neural network; (d) Reconstruction results based on CNN neural network.
Figure 8. The reconstruction results of the middle and long distance. (a) Speckle image corresponding to the target; (b) Reconstruction results based on DesUNet neural network; (c) Reconstruction results based on U-net neural network; (d) Reconstruction results based on CNN neural network.
Photonics 11 00887 g008
Figure 9. YOLO recognizes reconstructed target objects. (a) A speckle image; (b) The reconstructed image.
Figure 9. YOLO recognizes reconstructed target objects. (a) A speckle image; (b) The reconstructed image.
Photonics 11 00887 g009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, Y.; Zhou, C.; Dai, W. Long-Range Imaging through Scattering Media Using Deep Learning. Photonics 2024, 11, 887. https://doi.org/10.3390/photonics11090887

AMA Style

Jin Y, Zhou C, Dai W. Long-Range Imaging through Scattering Media Using Deep Learning. Photonics. 2024; 11(9):887. https://doi.org/10.3390/photonics11090887

Chicago/Turabian Style

Jin, Ying, Cangtao Zhou, and Wanjun Dai. 2024. "Long-Range Imaging through Scattering Media Using Deep Learning" Photonics 11, no. 9: 887. https://doi.org/10.3390/photonics11090887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop