Next Article in Journal
Simplified Beam Hardening Correction for Ultrafast X-ray CT Imaging of Binary Granular Mixtures
Previous Article in Journal
Gas Sensors Based on Semiconductor Metal Oxides Fabricated by Electrospinning: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single-Pixel Imaging Based on Deep Learning Enhanced Singular Value Decomposition

1
CAS Key Laboratory of Human-Machine Intelligence-Synergy Systems, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(10), 2963; https://doi.org/10.3390/s24102963
Submission received: 5 April 2024 / Revised: 27 April 2024 / Accepted: 2 May 2024 / Published: 7 May 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
We propose and demonstrate a single-pixel imaging method based on deep learning network enhanced singular value decomposition. The theoretical framework and the experimental implementation are elaborated and compared with the conventional methods based on Hadamard patterns or deep convolutional autoencoder network. Simulation and experimental results show that the proposed approach is capable of reconstructing images with better quality especially under a low sampling ratio down to 3.12%, or with fewer measurements or shorter acquisition time if the image quality is given. We further demonstrate that it has better anti-noise performance by introducing noises in the SPI systems, and we show that it has better generalizability by applying the systems to targets outside the training dataset. We expect that the developed method will find potential applications based on single-pixel imaging beyond the visible regime.

1. Introduction

Due to its function of adopting a single-pixel detector to collect the intensity of an object illuminated by a sequence of masked patterns, single-pixel imaging (SPI) is attractive in diverse applications such as ultrafast imaging [1,2], hyperspectral imaging [3,4], remote tracking [5,6], and three-dimensional imaging [7,8] for its low cost, high signal-to-noise ratio, and broadband operation [9,10,11]. The SPI techniques are especially important for wavelengths with expensive multi-pixel detectors, including terahertz [12,13,14,15], infrared [16,17,18,19], and X-ray [20,21,22,23].
In order to realize real-time SPI with high efficiency, many algorithms have been proposed or developed over the years. In 2008, Duarte et al. [24] proposed SPI based on a compressed sensing (CS) algorithm and reconstructed a 256 × 256 image using 1300 random patterns. However, under such a low sampling ratio, the reconstructed image suffers from relatively low quality. To improve the image quality while reducing the required acquisition time, several advanced methods based on Hadamard basis [25], Fourier basis [26], discrete cosine basis [27], and wavelet basis [28] have been developed, which leverage the completeness and orthogonality of basis patterns, as well as the sparse representation of natural images in the transform domain. Zhang et al. further compared, numerically and experimentally, the performances of Hadamard SPI (HSI) and Fourier SPI (FSI) [29], and demonstrated a 20,000 Hz projection rate using a DMD and captured 256 × 256 pixel dynamic scenes at a speed of 10 frames per second under the undersampling condition [30]. However, under the undersampling condition, these techniques truncate the spectrum by discarding the high-frequency parts, resulting in undesirable artifacts such as ringing in the reconstructed images [31].
Recently, deep learning networks have shown great potential in recovering high-quality images for SPI [32], especially when the sample ratio is exceptionally low. For example, Lyu et al. [33] developed a deep learning-based approach for ghost imaging (GI) and achieved significantly improved performance compared with the conventional CS method when the sampling ratio S R is down to 5%. He et al. [34] modified the convolutional neural network (CNN) so as to adapt it to GI, and achieved faster and more accurate image reconstruction than conventional GI when S R = 5 % . Higham et al. [35] proposed SPI based on a deep convolutional autoencoder network (DCAN) and recovered real-time 128 × 128 pixel video at 30 frames per second with a single-pixel camera at a sampling ratio of 2%. Rizvi et al. [31,36] developed a deep learning framework to enhance the imaging quality of four-step FSI and a fast image reconstruction framework based on DCAN for three-step FSI. They demonstrated that the proposed deep learning-based FSI outperforms the conventional FSI in terms of image quality, especially when the sampling ratio is down to 5–8%. Huang et al. [37] developed AuSamNet to optimize a sampling mask and reconstruct high-quality natural color images when the sampling ratio is as low as 7.5%. Zhu et al. [14] demonstrated an efficient terahertz SPI system incorporating deep learning networks and reduced the number of Hadamard patterns to 10% of the pixels while maintaining high image quality with acceptable signal-to-noise ratio (SNR). Stantchev et al. [38] demonstrated rapid terahertz SPI with a spatial light modulator and a convolutional neural network, and reduced S R to 10% while maintaining the image quality. Jiang et al. [39] proposed a novel SPI scheme for high-speed moving targets combined with a deep learning network and obtained reasonable reconstructions with a low sampling ratio of only 6%. Yao et al. [40] proposed a single-pixel classification method with deep learning for fast-moving objects and obtained feature information for classification with S R = 3.8 % . However, these data-driven networks suffer from problems such as generalizability and interpretability, which may prohibit their practical applications [41,42,43].
In order to address these problems, Wang et al. [44] proposed a physics-informed deep learning (PIDL) network and obtained better performance than other researchers using SPI algorithms at a low sampling ratio of 6.25%. We [15] demonstrated high-efficiency terahertz SPI system based on a PIDL network and reconstructed high-quality terahertz images with a significantly reduced number of measurements, corresponding to an ultra-low sampling ratio down to 1.56%, which is more efficient than the conventional HSI and FSI. However, these PIDL networks rely on GI or differential GI (DGI). It has been shown that the singular value decomposition (SVD) method can effectively enhance the fidelity and improve the imaging efficiency and quality of GI or DGI [45,46,47]. Quite recently, Cheng et al. [48] further combined the SVD-compressed GI with deep unfolding and numerically investigated its anti-noise performance at low sampling rates. However, in reports [45,46,48], the SVD algorithm was restricted to construct the encoding patterns, and the deep learning network in [48] was purely driven by data and thus also suffered from the above-mentioned problems such as generalizability and interoperability.
In this work, we propose a novel approach to realize efficient SPI based on deep learning-enhanced SVD (DLSVD). The theoretical framework and the experimental implementation are introduced, and the differences compared with conventional approaches are highlighted. We both numerically and experimentally demonstrate that the developed DLSVD outperforms the HSI, DCAN, and PIDL approaches in terms of the reconstructed image quality or the acquisition time. Remarkably, under the ultra-low sampling ratio down to 3.25%, the DLSVD can reconstruct images with improved quality over the HSI and the DCAN. The anti-noise performance and the generalization to objects of types outside the training dataset are also studied.

2. Theoretical Framework

Before we elaborate the DLSVD to be developed in this work, let us retrospect the classical HSI and DCAN schemes, which are adopted as references. For the HSI, one adopts Hadamard patterns to modulate the spatial intensity of optical beam illuminating the object, collects the transmitted signals with a single-pixel detector, and then reconstructs the image using the inverse Hadamard transform [29]. The Hadamard spectra comprise a series of Hadamard basis patterns in the spatial domain, P H u 0 , v 0 , which correspond to different Hadamard coefficients and which can be generated by
P H u 0 , v 0 ( x , y ) = 1 2 1 + H 1 δ ( u u 0 , v v 0 ) .
Here, H 1 { · } is the inverse Hadamard transform, δ ( u , v ) is the delta function, and ( u 0 , v 0 ) and ( u , v ) are Hadamard frequency points. For each measurement, the dot product of the object and the Hadamard basis pattern projected on it is recorded by a single-pixel detector
C H ( u 0 , v 0 ) = x = 0 N 1 y = 0 N 1 P H u 0 , v 0 ( x , y ) O ( x , y ) ,
where C H ( u 0 , v 0 ) is the coefficient of the Hadamard frequency point ( u 0 , v 0 ) , O ( x , y ) is the object of N × N pixels, and ( x , y ) represents the coordinate of the spatial domain.
By projecting a series of Hadamard basis patterns (of number M) onto the object, the cumulative signal can be expressed as [29]
D M × 1 = P M × N 2 O N 2 × 1 .
The sampling ratio is defined as S R M / N 2 . For the full sampling, M = N × N . We can use the reshape function to transform the one-dimensional vector D to the two-dimensional frequency coefficient C H ( u , v ) . The object’s image can be reconstructed by performing the inverse Hadamard transform R = H 1 C H ( u , v ) .
The DCAN approach, as illustrated in Figure 1a, adopts autoencoders (multilayer neural networks) to encode the object, and fully connected and convolutional layers to decode objects [35]. The encoding layer simulates the signal acquisition process of SPI. When encoding object O with patterns P of number M, signal S is also collected by a single-pixel detector. The encoding process with the i-th pattern can be represented as
( P i ) N × N O N × N = S i ,
where ⊙ denotes that the two matrices are bitwise multiplied by the elements and then added together. For M measurements, the encoding process can be represented by the matrix as
P O = S ,
where P is a matrix of M × N 2 , of which each row corresponds to a pattern of N × N , O is a matrix of N 2 × 1 , and S is a matrix of M × 1 .
The patterns P and the decoding network parameters θ are initialized randomly and then can be optimized using stochastic gradient descent to minimize the standard cost function that measures the Euclidean distance between the predicted and the desired outputs,
P , θ = min P P , θ Θ , R k O k 2 .
The DLSVD approach developed in this work is illustrated in Figure 1b. Similar to the PIDL, the proposed method also consists of an encoding layer, a decoding layer, and an enhancing layer. The encoding process can also be expressed as Equation (5), which is the same as that of the DCAN. Distinct from the literature on SPI involving the SVD, where the SVD method was used to construct the encoding patterns [45,46,48], here, the SVD algorithm is used for decoding instead. This is also distinct from the PIDL, of which the decoding makes use of the conventional DGI algorithm.
For full sampling with M = N 2 , the matrix P is a square matrix and can have an inverse, thus the target O can be reconstructed directly via
O = P 1 S .
For undersampling condition with M < N 2 , however, matrix P is not a square matrix but a singular matrix. For such a singular matrix, one can perform SVD to calculate its pseudo-inverse,
P = U Σ V T .
Here, U is an M × M orthogonal matrix whose column vectors are the eigenvectors of P P T , i.e., U U T = I . V T is an orthogonal matrix of N 2 × N 2 whose column vectors are the eigenvectors of P T P , i.e., V V T = I . Σ is a diagonal matrix of M × N 2 with elements on the diagonal being the singular values in descending order. In order to calculate the pseudo-inverse, we further truncate the singular values to the first K ones, where K M , and then obtain
O = V t Σ t 1 U t T S ,
where U t is a M × K matrix with orthogonal column vectors, Σ t is a K × K diagonal matrix, and V t T is a K × L matrix with orthogonal row vectors. O is further reshaped into the size of the input image, N × N .
We emphasize that for the developed DLSVD approach, the decoding layer adopts the SVD method, which is a noniterative and purely mathematical algorithm, to reconstruct the image. Therefore, compared with the DCAN, which adopts data-driven convolutional neural networks for image reconstruction, the developed DLSVD approach does not suffer from the interpretability problem, as demonstrated both numerically and experimentally in this work.
In order to reduce the noises in O , we add an enhancing layer composed of a convolutional neural network. Note that there is no restriction on the choice of neural network structure, although better results may be obtained by appropriately adapting the network structure. For instance, we utilize a U-net-like architecture, comprising five down-sampling layers and five up-sampling layers, designed to handle images of varying sizes [49,50]. The reconstructed image R after the enhancing layer can be then expressed as
R = U θ ( O ) ,
where θ represents the set of parameters used in the U-net.
Similar to the DCAN, the encoding patterns P and the parameters of the U-net enhancing layer, θ , should also be trained starting with random initialization and optimized with Equation (6). Note that for the DCAN, θ are the parameters of the decoding layer, whereas for the developed DLSVD, θ are the parameters of the enhancing layer.

3. Simulation Results

We first perform simulations to compare the performance of the developed DLSVD, the conventional HSI and the DCAN schemes, and the PIDL recently developed by Wang et al. [44]. Figure 2a shows the first three typical encoding patterns P among the M ones that are used in these SPI methods. For the HSI method, M measurements involve the first M encoding patterns within the well-established and well-ordered pattern pool, corresponding to the Hadamard spectra in the spatial domain [12]. In other words, for M = 2048 , the first 1024 patterns are exactly the same as those used for M = 1024 , and the first 512 patterns are the same as those used for M = 512 .
For the DCAN, the PIDL, and the DLSVD, however, the encoding patterns are first initialized randomly and then optimized through training, which should be performed independently for a specific measurement M. As a result, the encoding patterns of these approaches are different for different measurement times. We randomly select 45,000 images from the freely available ImageNet database [51] as the training set to optimize the encoding patterns and the parameters θ used in the DCAN, the PIDL and the DLSVD, and another disjoint 5000 images as the test set to test the trained network. All images are first uniformly resized to a resolution of 128 × 128 pixels and converted into gray-scale images, serving as the input for the networks. The learning rate is set to be 0.0002, and the number of epochs and the batch size are configured to 140 and 32, respectively. The training is carried out on an advanced gaming computer, boasting an Intel i9-12900k CPU (Intel, Santa Clara, CA, USA), 64 GB of RAM, and an NVIDIA RTX 3090 GPU (NVIDIA, Santa Clara, CA, USA).
Figure 2a further compares the first three typical trained encoding patterns of the DCAN, the PIDL, and the DLSVD with different measurements. We find that for the DCAN and the PIDL with different M, the trained encoding patterns all show random distributions of binary values of “1” and “ 1 ”. As M increases, the patterns have higher resolution, whereas for the DLSVD with different M, the first/second/third trained pattern composed of ternary values of “1“, “0”, and “ 1 ” shows pronounced similarities to its counterpart. This may correspond to the descending singular values, which are in the order of importance.
For different M, the trained encoding patterns of the DCAN, the PIDL, or the DLSVD are different since these patterns are first initialized randomly and then optimized through independently trained networks given the specific measurement number. Here, the networks for these approaches are defined and optimized following the details in the previous section. In other words, the optimized encoding patterns and the trained networks vary with M. This is distinct from the HSI, of which the first M patterns are taken from the well-defined pattern pool that is independent of M, as defined in Equation (1).
Figure 2b shows the reconstructed images O from the decoding layer using the SVD method and the reconstructed images R after the enhancing layer using the U-net structure. It is clear that the SVD decoding method is efficacious and that noises can be significantly removed by the U-net enhancing layer. In other words, the SVD-based SPI method can be greatly enhanced by the DL network. Visualization 1 shows the evolution of the first encoding patterns and the reconstructed images O and R after the decoding and enhancing layers, respectively, for M = 1024 .
Figure 3 shows the three typical ground truth (denoted as “GT”) images that are randomly taken from the test set within the ImageNet dataset and the reconstructed images by the HSI, the DCAN, the PIDL, and the developed DLSVD with different measurements of M = 512 , 1024, and 2048. Here, the traffic image subset of the ImageNet dataset is adopted for illustrating the comparison, although other target objects can also be applicable, as shown later in the experiments and on the generalizability. Simulation results show that the reconstructed images become clearer for all three methods as the number of measurements increases. For the traffic image of MotorVehicleOnly, the reconstructed images by the PIDL and DLSVD methods with M = 512 clearly show the outline of the bumper and lights of the car, whereas those by the HSI and the DCAN methods require 1024 measurements to manifest the outline, and 2048 measurements to show the bumper and lights relatively clearly. Remarkably, in the reconstructed images of the PIDL and DLSVD methods with M 1024 , one can find clear details of the bumper and lights. Similarly, for the other two traffic images of HonkingNeeded and StopAndGiveWay, the PIDL and the DLSVD methods with 512 measurements can reconstruct clear outlines, whereas the reconstructed images by the HSI are muddled, and those by the DCAN suffer from relatively high background noises. The reconstructed images by the PIDL and DLSVD methods are clean and clear when the measurement number increases to 1024. In contrast, for the HSI and the DCAN, this number should be at least 2048. Given the same number of measurements, in general, the DLSVD outperforms the PIDL, followed by the HSI and the DCAN.
To quantitatively assess the quality of the reconstructed images, we adopt two metrics, PSNR and SSIM [52], which can be calculated with
PSNR = 10 · log 10 255 2 MSE
and
SSIM = 2 ( u O u R + c 1 ) ( 2 σ OR + c 2 ) ( u O 2 + u R 2 + c 1 ) ( σ O 2 + σ R 2 + c 1 ) ,
respectively. Here, MSE is the mean square error defined as
MSE 1 N 2 x = 0 N 1 y = 0 N 1 O ( x , y ) R ( x , y ) 2 ,
where u O / R and σ O / R 2 represent the average value and the variance of the input image O or the reconstructed image R, respectively, σ O R denotes the covariance of the input image O and the reconstructed image R, and c 1 and c 2 are constants used to maintain stability.
In order to compare the anti-noise performance of the four SPI methods, we further add Gaussian white noises with variance of δ to the signal. The noise level is evaluated by the signal-to-noise ratio (SNR), which can be expressed as SNR = 10 · log 10 [ ( S S ¯ ) 2 ¯ / δ ] with the S ¯ signal perturbed by the Gaussian noises. Note that Poisson noise, which is also known as photon noise or shot noise and which is prominent in low-light conditions or when using imaging sensors with low light sensitivity, is neglected in this work since it is much weaker than the Gaussian white noises.
Figure 4 compares the SSIM and the PSNR of the images reconstructed by the four SPI methods with different numbers of measurements as functions of the noise level. Since the calculations were performed by averaging the reconstructed images of the test set of 5000 images, the standard deviations are denoted by the error bars. Results show that for each SPI approach, both the SSIM and the PSNR increase with the number of measurements regardless of the noise level. For noise levels of SNR 20 dB, the SSIM and the PSNR of all the four SPI methods are almost undisturbed. Given the number of measurements and under the same noise level, the DLSVD has slightly higher SSIM and PSNR than the PIDL, and larger SSIM and PSNR than the HSI and the DCAN methods. In other words, the developed DLSVD has the best anti-noise performance over the other three SPI methods.

4. Experimental Demonstrations

Encouraged by the simulation results, we build an experimental setup as illustrated in Figure 5 to perform experimental validations. White light beam from a 3 W LED is directed onto a Digital Micromirror Device (DMD) using Total Internal Reflection (TIR) prisms. The patterns on the DMD at a projection rate of 100 Hz are projected onto the target through a projection lens for achieving high signal-to-noise-rate images [30]. Here, we adopt the above three printed traffic images shown in Figure 3 as the target. The reflected light from the target is captured by a single-pixel avalanche photodiode (Thorlabs APD410A2/M, Thorlabs, Newton, NJ, USA). Data are collected with the data acquisition board of National Instruments (USB6361). The DMD consists of 1024 × 768 micro-mirrors, each of which has a pitch size of 13.68 μ m and acts as a binary reflector projecting two states: “0” and “1”, corresponding to black and white, respectively. The encoding patterns used in experiments for the four SPI methods are the same as those used in simulations, as shown in Figure 2a. In order to realize the encoding pattern element “ 1 ” shown in Figure 2a, we employ a differential measurement method, which splits the original encoding mask with “ 1 ” elements into two masks containing only “0” and “1” elements. Therefore, for each measurement, these masks are sampled individually, and the signals collected by the single-pixel detector are subtracted from each other for further processing.
Figure 6 presents the reconstructed images by the three SPI methods with different numbers of measurements. Comparing Figure 3 and Figure 6, we find that in general, the experimental results agree well with the simulations: for all the four SPI methods, the reconstructed image quality improves as M increases; given the number of measurements, the developed DLSVD slightly outperforms the PIDL, which is followed by the HSI and the DCAN. Strikingly, high-quality reconstructed images can be obtained by the PIDL and the developed DLSVD even when M = 512 , corresponding to an ultra-low sampling ratio of S R = 3.12 % .
To quantitatively evaluate the quality of the experimentally reconstructed images, we calculate the PSNR and the SSIM by choosing the reconstructed images by the HSI at the full sampling rate of S R = 100 % , or equivalently M = 128 × 128 , as the ground truth. The results summarized in Table 1 show that the larger the number of measurements, the higher the PSNR and the SSIM all for the four SPI methods. For most cases, the proposed DLSVD method has the highest PSNR and SSIM compared with the other three SPI methods. For a few cases, the PSNR and the SSIM of the images reconstructed by the PIDL are slightly better than those recobstructed by the DLSVD.
Specifically, for M = 512 , the PSNR of the DLSVD is slightly larger than that of the PIDL for two images, but it is ∼1–2 dB larger than those of the HSI and the DCAN for all the three images; the SSIM of the DLSVD is slightly larger than that of the PIDL for the image of MotorVehicleOnly, but is much larger than those of the DCAN and the HSI for all the three images. As the number of measurements increases to 1024 or 2048, the differences between the PSNRs of the DLSVD and those of the other three methods further increase to ∼2–3 dB or ∼3–4 dB. On the SSIM, the DLSVD outperforms the PIDL for two images when M = 1024 and for all three images when M = 2048 . All these quantitative metrics are consistent with the intuition from the reconstructed images by experiments (in Figure 6) as well as by simulations (in Figure 3). To offer a more comprehensive evaluation of DLSVD, we include cross-dimensional comparisons using different measurement settings. When M = 1024 , DLSVD achieves a mean PSNR of 24.33 dB and an SSIM of 0.7351. In contrast, when M = 2048 , HSI yields 22.16 dB and 0.7080, DCAN scores 22.21 dB and 0.69982, and PIDL achieves 23.56 dB and 0.7575. Notably, DLSVD outperforms HSI and DCAN, slightly surpassing PIDL in performance. In other words, we experimentally verify that, given the number of measurements, the proposed DLSVD system can reconstruct images of better quality than the HSI, the DCAN, and the PIDL methods.
The improvement in the quality of the reconstructed images is equivalent to the reduction in acquisition time, which equals the sum of the sampling time and the reconstruction time. In our experiments, all the four SPI methods share the same sampling time per pattern, which is 0.001 s. The reconstruction time is 0.02 s for the HSI, 0.004 s for the DCAN, 0.009 s for the PIDL, and 0.006 s for the DLSVD. Compared with the HSI or the DCAN, the latter of which has the shortest reconstruction time, the DLSVD can reduce the acquisition time by 50% due to half measurements required to achieve the same reconstructed image quality. Compared with the PIDL, which has comparable reconstructed image quality under the same M, or equivalently using the same sampling time, the DLSVD can reduce the reconstruction time by 1/3. Therefore, given the reconstructed image quality, the DLSVD is more efficient since it requires fewer measurements than the HSI and the DCAN, or has faster reconstruction time than the PIDL.

5. Extension to Other Objects

Although till now we have considered objects only from the ImageNet dataset, the concept and the conclusions are general, and the optimized patterns and the trained networks can be applied to other objects outside the ImageNet dataset. As an illustration, we printed four other types of objects, which are letters, cartoon patterns, and numbers. In these SPI experiments, the patterns, the networks, and the setups are the same as those used in the above experiments.
Figure 7 plots the reconstructed images by the four SPI methods. It is clear that the PIDL or the DLSVD outperforms the HSI and the DCAN methods regardless of the number of measurements. Particularly for an ultra-small number of measurements ( M = 512 ) which corresponds to an ultra-low sampling ratio of S R = 3.12 % , the images reconstructed by both the PIDL and the DLSVD have sharp edges and large SNR, whereas those by the DCAN are slightly blurred and those by the HSI suffer from relatively smaller SNR.
We further extracted the PSNRs and the SSIMs of these reconstructed images, and the results are shown in Table 2. We found that for all the four SPI systems, the quality of reconstructed images improves with M, that the PIDL or the DLSVD have the largest PSNRs and SSIMs compared with the HSI and the DCAN for different M, and that the DCAN is slightly better than the HSI. These results are consistent with those obtained with targets chosen from the ImageNet dataset (Figure 6). The exception is that the PIDL is slightly better than the DLSVD for M = 512 or 1024, whereas the DLSVD outperforms the PIDL for M = 2048 .
Therefore, our experiment results showed that both the PIDL and the developed DLSVD SPI systems have better generalizability than the conventional HSI and DCAN systems. This is because the DL networks in both the PIDL and the DLSVD are used in the enhancing layer rather than in the decoding layer. In contrast, the DCAN takes advantage of the DL networks for decoding.
Finally, we illustrated the generalizability of the developed DLSVD with the trained patterns and networks by reconstructing complex target objects such as Cameraman. Simulation results in Figure 8 show that the developed DLSVD outperforms the other three SPI methods in terms of the image quality evaluated by the SSIM and the PNSR regardless of the number of measurements. This further validates that the developed DLSVD has the best performance in generalizability.

6. Conclusions

In conclusion, we proposed and demonstrated an efficient single-pixel imaging method based on DLSVD. We elaborated the theoretical framework and experimental implementation of the developed DLSVD system by comparison with the HSI, DCAN, and PIDL methods, highlighting their similarities and differences. Thanks to the encoding patterns that are optimized by the trained network, both the simulation and experimental results showed that, compared with the other three SPI methods, the developed DLSVD is more efficient in terms of the number of measurements or the acquisition time required for clear image reconstruction, and it is more robust to noises. Strikingly, even when the sampling ratio is down to 3.12%, which corresponds to 512 measurements for 128 × 128 pixel images, the experimentally extracted PSNRs are larger than 22 and the SSIMs are greater than 0.61 for three typical targets from the ImageNet dataset. We also demonstrated that the DLSVD outperforms the conventional HSI and DCAN systems in terms of anti-noise performance by introducing noises in the systems and of the generalizability by directly applying the SPI systems to targets outside the ImageNet dataset. Therefore, we expect this work to advance the development of SPI techniques in a variety of applications, ranging from remote sensing to microscopy beyond the visible regime.

Author Contributions

Conceptualization, R.S., Y.L. and G.L.; methodology, Y.D. and R.S.; software, R.S. and Y.D.; validation, Y.D. and R.S.; formal analysis, Y.D., R.S., W.L., Y.L. and G.L.; investigation, Y.D., R.S. and G.L.; resources, R.S., W.L., Y.L. and G.L.; data curation, Y.D. and R.S.; writing—original draft preparation, Y.D., R.S. and G.L.; writing—review and editing, Y.D., R.S., W.L., Y.L. and G.L.; visualization, Y.D., R.S. and W.L.; supervision, Y.L. and G.L.; project administration, R.S., Y.L. and G.L.; funding acquisition, R.S., W.L. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (62105354); the National Key Research and Development Program of China (2019YFB1312404); the SIAT Innovation Program for Excellent Young Researchers; the Major Instrumentation Development Program of the Chinese Academy of Sciences (ZDKYYQ20220008).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Acknowledgments

The authors thank the coach and players of the team who participated in the study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xu, Z.H.; Chen, W.; Penuelas, J.; Padgett, M.; Sun, M.J. 1000 fps computational ghost imaging using LED-based structured illumination. Opt. Express 2018, 26, 2427–2434. [Google Scholar] [CrossRef] [PubMed]
  2. Zhao, W.; Chen, H.; Yuan, Y.; Zheng, H.; Liu, J.; Xu, Z.; Zhou, Y. Ultrahigh-Speed Color Imaging with Single-Pixel Detectors at Low Light Level. Phys. Rev. Appl. 2019, 12, 034049. [Google Scholar] [CrossRef]
  3. Jin, S.; Hui, W.; Wang, Y.; Huang, K.; Shi, Q.; Ying, C.; Liu, D.; Ye, Q.; Zhou, W.; Tian, J. Hyperspectral imaging using the single-pixel Fourier transform technique. Sci. Rep. 2017, 7, 45209. [Google Scholar] [CrossRef] [PubMed]
  4. Amiot, C.; Ryczkowski, P.; Friberg, A.T.; Dudley, J.M.; Genty, G. Supercontinuum spectral-domain ghost imaging. Opt. Lett. 2018, 43, 5025–5028. [Google Scholar] [CrossRef] [PubMed]
  5. Ma, S.; Liu, Z.; Wang, C.; Hu, C.; Li, E.; Gong, W.; Tong, Z.; Wu, J.; Shen, X.; Han, S. Ghost imaging LiDAR via sparsity constraints using push-broom scanning. Opt. Express 2019, 27, 13219–13228. [Google Scholar] [CrossRef]
  6. Sun, S.; Lin, H.; Xu, Y.; Gu, J.; Liu, W. Tracking and imaging of moving objects with temporal intensity difference correlation. Opt. Express 2019, 27, 27851–27861. [Google Scholar] [CrossRef] [PubMed]
  7. Sun, B.; Edgar, M.P.; Bowman, R.; Vittert, L.E.; Welsh, S.; Bowman, A.; Padgett, M.J. 3D Computational Imaging with Single-Pixel Detectors. Science 2013, 340, 844–847. [Google Scholar] [CrossRef] [PubMed]
  8. Yu, H.; Li, E.; Gong, W.; Han, S. Structured image reconstruction for three-dimensional ghost imaging lidar. Opt. Express 2015, 23, 14541–14551. [Google Scholar] [CrossRef]
  9. Gibson, G.M.; Johnson, S.D.; Padgett, M.J. Single-pixel imaging 12 years on: A review. Opt. Express 2020, 28, 28190–28208. [Google Scholar] [CrossRef]
  10. Edgar, M.P.; Gibson, G.M.; Padgett, M.J. Principles and prospects for single-pixel imaging. Nat. Photonics 2019, 13, 13–20. [Google Scholar] [CrossRef]
  11. Sun, M.J.; Zhang, J.M. Single-Pixel Imaging and Its Application in Three-Dimensional Reconstruction: A Brief Review. Sensors 2019, 19, 732. [Google Scholar] [CrossRef] [PubMed]
  12. She, R.; Liu, W.; Lu, Y.; Zhou, Z.; Li, G. Fourier single-pixel imaging in the terahertz regime. Appl. Phys. Lett. 2019, 115, 021101. [Google Scholar] [CrossRef]
  13. She, R.B.; Liu, W.Q.; Wei, G.L.; Lu, Y.F.; Li, G.Y. Terahertz Single-Pixel Imaging Improved by Using Silicon Wafer with SiO2 Passivation. Appl. Sci. 2020, 10, 2427. [Google Scholar] [CrossRef]
  14. Zhu, Y.; She, R.; Liu, W.; Lu, Y.; Li, G. Deep Learning Optimized Terahertz Single-Pixel Imaging. IEEE Trans. Terahertz Sci. Technol. 2022, 12, 165–172. [Google Scholar] [CrossRef]
  15. Deng, Y.; She, R.; Liu, W.; Lu, Y.; Li, G. High-efficiency terahertz single-pixel imaging based on a physics-enhanced network. Opt. Express 2023, 31, 10273–10286. [Google Scholar] [CrossRef] [PubMed]
  16. Gibson, G.M.; Sun, B.; Edgar, M.P.; Phillips, D.B.; Hempler, N.; Maker, G.T.; Malcolm, G.P.A.; Padgett, M.J. Real-time imaging of methane gas leaks using a single-pixel camera. Opt. Express 2017, 25, 2998–3005. [Google Scholar] [CrossRef]
  17. Zeng, B.; Huang, Z.; Singh, A.; Yao, Y.; Azad, A.K.; Mohite, A.D.; Taylor, A.J.; Smith, D.R.; Chen, H.T. Hybrid graphene metasurfaces for high-speed mid-infrared light modulation and single-pixel imaging. Light Sci. Appl. 2018, 7, 51. [Google Scholar] [CrossRef] [PubMed]
  18. Gattinger, P.; Kilgus, J.; Zorin, I.; Langer, G.; Nikzad-Langerodi, R.; Rankl, C.; Gröschl, M.; Brandstetter, M. Broadband near-infrared hyperspectral single pixel imaging for chemical characterization. Opt. Express 2019, 27, 12666–12672. [Google Scholar] [CrossRef] [PubMed]
  19. Wang, Y.; Huang, K.; Fang, J.; Yan, M.; Wu, E.; Zeng, H. Mid-infrared single-pixel imaging at the single-photon level. Nat. Commun. 2023, 14, 1073. [Google Scholar] [CrossRef] [PubMed]
  20. Schori, A.; Shwartz, S. X-ray ghost imaging with a laboratory source. Opt. Express 2017, 25, 14822–14828. [Google Scholar] [CrossRef] [PubMed]
  21. Klein, Y.; Schori, A.; Dolbnya, I.P.; Sawhney, K.; Shwartz, S. X-ray computational ghost imaging with single-pixel detector. Opt. Express 2019, 27, 3284–3293. [Google Scholar] [CrossRef]
  22. He, Y.H.; Zhang, A.X.; Li, M.F.; Huang, Y.Y.; Quan, B.G.; Li, D.Z.; Wu, L.A.; Chen, L.M. High-resolution sub-sampling incoherent x-ray imaging with a single-pixel detector. APL Photonics 2020, 5, 056102. [Google Scholar] [CrossRef]
  23. Olbinado, M.P.; Paganin, D.M.; Cheng, Y.; Rack, A. X-ray phase-contrast ghost imaging using a single-pixel camera. Optica 2021, 8, 1538–1544. [Google Scholar] [CrossRef]
  24. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef]
  25. Sun, M.J.; Meng, L.T.; Edgar, M.P.; Padgett, M.J.; Radwell, N. A Russian Dolls ordering of the Hadamard basis for compressive single-pixel imaging. Sci. Rep. 2017, 7, 3464. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, Z.; Ma, X.; Zhong, J. Single-pixel imaging by means of Fourier spectrum acquisition. Nat. Commun. 2015, 6, 6225. [Google Scholar] [CrossRef] [PubMed]
  27. Chen, Y.; Liu, S.; Yao, X.R.; Zhao, Q.; Liu, X.F.; Liu, B.; Zhai, G.J. Discrete cosine single-pixel microscopic compressive imaging via fast binary modulation. Opt. Commun. 2020, 454, 124512. [Google Scholar] [CrossRef]
  28. Rousset, F.; Ducros, N.; Farina, A.; Valentini, G.; Andrea, C.D.; Peurin, F. Adaptive Basis Scan by Wavelet Prediction for Single-Pixel Imaging. IEEE Trans. Comput. Imaging 2017, 3, 36–46. [Google Scholar] [CrossRef]
  29. Zhang, Z.; Wang, X.; Zheng, G.; Zhong, J. Hadamard single-pixel imaging versus Fourier single-pixel imaging. Opt. Express 2017, 25, 19619–19639. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Wang, X.; Zheng, G.; Zhong, J. Fast Fourier single-pixel imaging via binary illumination. Sci. Rep. 2017, 7, 12029. [Google Scholar] [CrossRef] [PubMed]
  31. Rizvi, S.; Cao, J.; Zhang, K.Y.; Hao, Q. Deringing and denoising in extremely under-sampled Fourier single pixel imaging. Opt. Express 2020, 28, 7360–7374. [Google Scholar] [CrossRef] [PubMed]
  32. Jacome, R.; Gomez, P.; Arguello, H. Middle output regularized end-to-end optimization for computational imaging. Optica 2023, 10, 1421–1431. [Google Scholar] [CrossRef]
  33. Lyu, M.; Wang, W.; Wang, H.; Wang, H.; Li, G.; Chen, N.; Situ, G. Deep-learning-based ghost imaging. Sci. Rep. 2017, 7, 17865. [Google Scholar] [CrossRef]
  34. He, Y.; Wang, G.; Dong, G.; Zhu, S.; Chen, H.; Zhang, A.; Xu, Z. Ghost Imaging Based on Deep Learning. Sci. Rep. 2018, 8, 6469. [Google Scholar] [CrossRef]
  35. Higham, C.F.; Murray-Smith, R.; Padgett, M.J.; Edgar, M.P. Deep learning for real-time single-pixel video. Sci. Rep. 2018, 8, 2369. [Google Scholar] [CrossRef] [PubMed]
  36. Rizvi, S.; Cao, J.; Zhang, K.; Hao, Q. Improving Imaging Quality of Real-time Fourier Single-pixel Imaging via Deep learning. Sensors 2019, 19, 4190. [Google Scholar] [CrossRef] [PubMed]
  37. Huang, W.; Wang, F.; Zhang, X.; Jin, Y.; Situ, G. Learning-based adaptive under-sampling for Fourier single-pixel imaging. Opt. Lett. 2023, 48, 2985–2988. [Google Scholar] [CrossRef]
  38. Stantchev, R.I.; Li, K.; Pickwell-MacPherson, E. Rapid Imaging of Pulsed Terahertz Radiation with Spatial Light Modulators and Neural Networks. ACS Photonics 2021, 8, 3150–3155. [Google Scholar] [CrossRef]
  39. Jiang, W.; Li, X.; Peng, X.; Sun, B. Imaging high-speed moving targets with a single-pixel detector. Opt. Express 2020, 28, 7889–7897. [Google Scholar] [CrossRef]
  40. Yao, M.; Zheng, S.; Hu, Y.; Zhang, Z.; Peng, J.; Zhong, J. Single-Pixel Moving Object Classification with Differential Measuring in Transform Domain and Deep Learning. Photonics 2022, 9, 202. [Google Scholar] [CrossRef]
  41. Wang, F.; Bian, Y.; Wang, H.; Lyu, M.; Pedrini, G.; Osten, W.; Barbastathis, G.; Situ, G. Phase imaging with an untrained neural network. Light Sci. Appl. 2020, 9, 77. [Google Scholar] [CrossRef] [PubMed]
  42. Mur, A.L.; Leclerc, P.; Peyrin, F.; Ducros, N. Single-pixel image reconstruction from experimental data using neural networks. Opt. Express 2021, 29, 17097–17110. [Google Scholar] [CrossRef] [PubMed]
  43. Karniadakis, G.E.; Kevrekidis, I.G.; Lu, L.; Perdikaris, P.; Wang, S.; Yang, L. Physics-informed machine learning. Nat. Rev. Phys. 2021, 3, 422–440. [Google Scholar] [CrossRef]
  44. Wang, F.; Wang, C.; Deng, C.; Han, S.; Situ, G. Single-pixel imaging using physics enhanced deep learning. Photonics Res. 2022, 10, 104–110. [Google Scholar] [CrossRef]
  45. Zhang, X.; Meng, X.; Yang, X.; Wang, Y.; Yin, Y.; Li, X.; Peng, X.; He, W.; Dong, G.; Chen, H. Singular value decomposition ghost imaging. Opt. Express 2018, 26, 12948–12958. [Google Scholar] [CrossRef] [PubMed]
  46. Zhang, C.; Tang, J.; Zhou, J.; Wei, S. Singular value decomposition compressed ghost imaging. Appl. Phys. B-Lasers Opt. 2022, 128, 47. [Google Scholar] [CrossRef]
  47. Chen, L.Y.; Wang, C.; Xiao, X.Y.; Ren, C.; Zhang, D.J.; Li, Z.; Cao, D.Z. Denoising in SVD-based ghost imaging. Opt. Express 2022, 30, 6248–6257. [Google Scholar] [CrossRef] [PubMed]
  48. Zhang, C.; Zhou, J.; Tang, J.; Wu, F.; Cheng, H.; Wei, S. Deep unfolding for singular value decomposition compressed ghost imaging. Appl. Phys. B-Lasers Opt. 2022, 128, 185. [Google Scholar] [CrossRef]
  49. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  50. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar]
  51. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  52. Ye, Z.; Qiu, P.; Wang, H.; Xiong, J.; Wang, K. Image watermarking and fusion based on Fourier single-pixel imaging with weighed light source. Opt. Express 2019, 27, 36505–36523. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Schematic of the DCAN adapted from [35], which includes the encoding and decoding layers. (b) Diagram of the developed DLSVD framework, which consists of the encoding, decoding, and enhancing layers.
Figure 1. (a) Schematic of the DCAN adapted from [35], which includes the encoding and decoding layers. (b) Diagram of the developed DLSVD framework, which consists of the encoding, decoding, and enhancing layers.
Sensors 24 02963 g001
Figure 2. (a) First three typical encoding patterns for the four SPI methods with different measurements of M = 512 , 1024, and 2048. (b) Reconstructed images by the developed DLSVD with M = 512 , 1024, and 2048.
Figure 2. (a) First three typical encoding patterns for the four SPI methods with different measurements of M = 512 , 1024, and 2048. (b) Reconstructed images by the developed DLSVD with M = 512 , 1024, and 2048.
Sensors 24 02963 g002
Figure 3. Three typical ground truth (GT) images taken from the ImageNet dataset, and reconstructed images by the four SPI methods with different measurements of M = 512 , 1024, and 2048.
Figure 3. Three typical ground truth (GT) images taken from the ImageNet dataset, and reconstructed images by the four SPI methods with different measurements of M = 512 , 1024, and 2048.
Sensors 24 02963 g003
Figure 4. (a) Extracted SSIMs and (b) PSNRs of the reconstructed images by the four SPI methods with different numbers of measurements as functions of the ambient noise. Error bars indicate standard deviations of the test set.
Figure 4. (a) Extracted SSIMs and (b) PSNRs of the reconstructed images by the four SPI methods with different numbers of measurements as functions of the ambient noise. Error bars indicate standard deviations of the test set.
Sensors 24 02963 g004
Figure 5. Schematic of the experimental setup.
Figure 5. Schematic of the experimental setup.
Sensors 24 02963 g005
Figure 6. Ground truth (GT) images taken from the reconstructed ones by the HSI method with full sampling ratio ( M = N 2 ) and experimentally reconstructed images by the four SPI methods with different measurements ( M = 512 , 1024, 2048).
Figure 6. Ground truth (GT) images taken from the reconstructed ones by the HSI method with full sampling ratio ( M = N 2 ) and experimentally reconstructed images by the four SPI methods with different measurements ( M = 512 , 1024, 2048).
Sensors 24 02963 g006
Figure 7. Similar to Figure 6 but with different images.
Figure 7. Similar to Figure 6 but with different images.
Sensors 24 02963 g007
Figure 8. Ground truth (GT) images taken from the reconstructed ones by the HSI method with full sampling ratio ( M = N 2 ), and numerically reconstructed images by the four SPI methods with different measurements of M = 512 , 1024, 2048. The values of SSIM and PSNR are extracted for each reconstructed image.
Figure 8. Ground truth (GT) images taken from the reconstructed ones by the HSI method with full sampling ratio ( M = N 2 ), and numerically reconstructed images by the four SPI methods with different measurements of M = 512 , 1024, 2048. The values of SSIM and PSNR are extracted for each reconstructed image.
Sensors 24 02963 g008
Table 1. Extracted PSNR and SSIM of the reconstructed images by the four SPI methods with different numbers of measurements. The largest values are highlighted in bold.
Table 1. Extracted PSNR and SSIM of the reconstructed images by the four SPI methods with different numbers of measurements. The largest values are highlighted in bold.
MeasurementMethodMotorVehicleOnlyHonkingNeededStopAndGiveWay
PSNRSSIMPSNRSSIMPSNRSSIM
512HSI20.12530.5328320.67220.5778019.77860.61145
DCAN21.07880.5848919.57970.6409421.36270.71429
PIDL22.01910.6118124.23390.6964921.86390.74317
DLSVD22.90140.6163124.17650.6864722.10560.73374
1024HSI20.33220.5992221.43090.6724721.30500.70870
DCAN22.08330.6074320.50350.6659221.17870.72877
PIDL20.51110.6405524.18970.7195323.97610.80488
DLSVD23.20490.6724926.02260.7282523.77350.80481
2048HSI20.99760.6559922.25910.7113523.22660.75675
DCAN21.05480.6471124.05830.6854321.52860.76694
PIDL21.89620.6945524.72080.7463724.07130.83182
DLSVD25.22500.7234026.42190.7607525.83880.84479
Bold number denotes the best performance value among methods.
Table 2. Two indicators of experimentally reconstructed images by the four SPI methods.
Table 2. Two indicators of experimentally reconstructed images by the four SPI methods.
MeasurementMethodLettersCartoonNumbers
PSNRSSIMPSNRSSIMPSNRSSIM
512HSI17.58260.5613116.91970.5938816.00130.47396
DCAN19.77880.6439517.51720.6407318.16960.63210
PIDL22.86700.8141421.20270.7390622.86700.81414
DLSVD21.20670.7815519.87130.7583319.55160.74773
1024HSI18.30330.6705417.58070.6297517.86980.63145
DCAN18.81460.6667019.63560.6876419.59110.67321
PIDL21.54320.8097221.39740.7845723.81220.82897
DLSVD21.95170.8043221.07320.7660623.02440.82127
2048HSI20.10110.7124717.85950.6896018.40020.69709
DCAN15.12680.6527918.98680.7049219.63540.70037
PIDL24.60530.8395623.53650.8024124.63240.83752
DLSVD26.16520.8415424.12260.8590125.12780.86156
Bold number denotes the best performance value among methods.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, Y.; She, R.; Liu, W.; Lu, Y.; Li, G. Single-Pixel Imaging Based on Deep Learning Enhanced Singular Value Decomposition. Sensors 2024, 24, 2963. https://doi.org/10.3390/s24102963

AMA Style

Deng Y, She R, Liu W, Lu Y, Li G. Single-Pixel Imaging Based on Deep Learning Enhanced Singular Value Decomposition. Sensors. 2024; 24(10):2963. https://doi.org/10.3390/s24102963

Chicago/Turabian Style

Deng, Youquan, Rongbin She, Wenquan Liu, Yuanfu Lu, and Guangyuan Li. 2024. "Single-Pixel Imaging Based on Deep Learning Enhanced Singular Value Decomposition" Sensors 24, no. 10: 2963. https://doi.org/10.3390/s24102963

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop