Next Article in Journal
Prototype Optical Bionic Microphone with a Dual-Channel Mach–Zehnder Interferometric Transducer
Previous Article in Journal
Detection and Classification of Cotton Foreign Fibers Based on Polarization Imaging and Improved YOLOv5
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

On-Chip Compressive Sensing with a Single-Photon Avalanche Diode Array

School of Electrical Science and Engineering, Nanjing University, Nanjing 210023, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(9), 4417; https://doi.org/10.3390/s23094417
Submission received: 2 April 2023 / Revised: 21 April 2023 / Accepted: 28 April 2023 / Published: 30 April 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Single-photon avalanche diodes (SPADs) are novel image sensors that record photons at extremely high sensitivity. To reduce both the required sensor area for readout circuits and the data throughput for SPAD array, in this paper, we propose a snapshot compressive sensing single-photon avalanche diode (CS-SPAD) sensor which can realize on-chip snapshot-type spatial compressive imaging in a compact form. Taking advantage of the digital counting nature of SPAD sensing, we propose to design the circuit connection between the sensing unit and the readout electronics for compressive sensing. To process the compressively sensed data, we propose a convolution neural-network-based algorithm dubbed CSSPAD-Net which could realize both high-fidelity scene reconstruction and classification. To demonstrate our method, we design and fabricate a CS-SPAD sensor chip, build a prototype imaging system, and demonstrate the proposed on-chip snapshot compressive sensing method on the MINIST dataset and real handwritten digital images, with both qualitative and quantitative results.

1. Introduction

Due to its digitized sensing nature of light and high sensitivity, single-photon avalanche diode (SPAD) imagers have been applied in a variety of applications, such as low light imaging, high dynamic range imaging, etc. [1,2,3]. Realizing a two-dimensional SPAD sensor with a large pixel number is a long-pursued goal, which could be readily applied for computer vision tasks [4,5,6,7,8]. However, due to the working principle of SPAD, the requirement of the self-contained controlling circuit, in-pixel or periphery data acquisition, or storage memory become the bottleneck for fabricating the SPAD arrays with a large pixel number [9]. Furthermore, the growing requirement of bandwidth with the increase of the pixel number poses another challenge. To realize format SPAD array, either column-wise shared periphery circuits [10] or 3D stacking technology [11] is used, facing the trade-off between the readout speed and high cost, with non-reduced data bandwidth. Therefore, how to efficiently realize a large-format SPAD array is still an unresolved problem. Since 2006, the theory of compressed sensing (CS) [12,13,14] has been proposed for efficient data sampling and high-fidelity sparse recovery. Implementing the idea of CS on the CMOS (complementary metal–oxide–semiconductor) sensor chip has been recently proposed [15,16,17,18,19,20,21] and shows promising data throughput reduction on conventional sensing techniques.
In this paper, to realize efficient SPAD array sensing, we introduce compressed imaging technology to the design of the SPAD array. Our CS-SPAD array chip could implement compressive sensing in a snapshot way and reduce the total bandwidth of imaging data to 25%. Besides, the required sensor area for readout electronics with memory is reduced by 25%.
The entire pipeline of the proposed scheme is shown in Figure 1. Specifically, we propose to realize compressive sensing on a chip with a 32 × 32 SPAD array. Through introducing different electronic connections between the readout electronics and the sensing pixels, photons collected by different combinations of sensor pixels can be captured in a summation way. The connections are set according to the compressive sensing matrix and the measurement of each readout electronics corresponds to a CS measurement. Furthermore, to retrieve information from the compressively sensed data of the CS-SPAD imaging sensor, we propose a CSSPAD-Net which could realize both image recovery and image classification based on the captured data of the CS-SPAD array. To demonstrate the proposed method, we design and fabricate the CS-SPAD sensor array, build a prototype imaging system based on the chip, and realize both image recovery and classification based on the output of the sensor, with high fidelity. In all, with the proposed CS-SPAD sensor, we could realize SPAD array sensing efficiently, i.e., with 25% data throughput and 25% periphery electronics required for two-dimensional sensing. The performance of imaging and classification of the CS-SPAD imaging system is demonstrated on the MNIST [22] dataset and real handwritten digital images, both quantitatively and qualitatively.

2. Related Work

Large-format SPAD imaging arrays have been a long-pursued goal [9,10,11,23,24,25], with wide application in low-light imaging [1,26], high-dynamic-range imaging [2,3], 3D imaging [1,27], etc. To realize the photon-resolving working principle of SPAD, the peripheral circuit, including the quenching circuit, readout electronics, etc., for generating the photon counting results are required and occupy a large area of the SPAD sensor, preventing the implementation of a large-format SPAD imager. Commonly, to realize large-format SPAD arrays, the peripheral circuits or counting electronics with memory are commonly shared among columns or rows of pixels [10], facing a trade-off between the readout speed and required sensor area. To realize large-format SPAD imaging without sharing the readout electronics, 3D stacking technology is introduced and large-format SPAD imaging can be realized, at the expense of high cost [11,28]. Furthermore, the requirement of data bandwidth are not reduced, becoming a potential bottleneck for large-format SPAD arrays. Thus, how to realize a SPAD imager with an efficient sensor area is still an open problem.
Compressive sensing [12], firstly proposed in 2006, provides an elegant solution to reduce the required data transmission bandwidth through sampling the scene information with a sub-Nyquist sampling rate and compressive reconstruction, taking full advantage of nature image redundancy. Mathematically, the scene redundancy is modeled with sparsity in the image transform domain, e.g., Fourier or wavelet transform domain [29], learned transform domain [30,31], etc. Through restricting the sparseness of the signal in the transform domain, the original signal can be recovered with high fidelity from the compressive measurement. Based on the compressive sensing theory, efficient imaging technology based on an optical system is proposed, such as single-pixel imaging [32], ghost imaging [33], low-light imaging [34] and mid-infrared imaging [35], which could largely reduce the required data transmission bandwidth.
Sun et al. [36] proposed to build a discrete micromirror device modulation-based compressive SPAD imaging system to realize a high-resolution SPAD imager. However, the requirement of a spatial light modulator leads to a large increase of the imaging complexity. Besides, multiple measurements are required for different compressed codes, which largely restricts the speed of compression acquisition and prevents real-time imaging.
In order to avoid complex optical systems, some on-chip compressed sensing schemes have been proposed in conventional imaging sensors, which could realize efficient data reading and reduced data bandwidth. Conventional CMOS image sensor converts light intensity into electrical signals for each pixel individually, while CS CMOS image sensors only sample a small set of random pixel summations [15,16,17,18,19,20,21], which can reduce the size of output data, analog to digital conversion (ADC) operations and the sensor power consumption.
In this paper, we propose to implement compressive sensing on the SPAD sensor array, which could realize compressive sensing on the chip and help to reduce the data throughput and required sensor area for data reading and memory. Specifically, we design the electrical circuits between the sensing unit and the readout electronics to realize the compressive sensing process. Each set of readout electronics is designed to count the pixels in a local unit and integrate the process of data compression into the chip through the local coupling of pixels on the chip, which reduces the sensor area required for data storage and data throughput to 25%.
To demonstrate our methods, we propose CSSPAD-Net to process the captured data by the CS SPAD chip, fabricate and tape out the CS-SPAD image sensor, and build a prototype imaging system. The effectiveness and efficiency of the proposed CS-SPAD sensor are demonstrated, both quantitatively and qualitatively, on the MNIST dataset [22] or real handwritten digital images. We will introduce the details of our method in the methods section.

3. Methods

In order to realize efficient SPAD array sensing, we design a novel compressive sensing SPAD array which can directly record the compressively sensed data. In the decoding process, we propose a neural network designed to directly process the compressively sensed data, which can reconstruct the scene and realize classification. This section will introduce our methods, including the proposed compressive imaging chip design and information processing network architecture for compressed data.

3.1. Snapshot Compressed Imaging Chip

3.1.1. Basic Compressed Coding Unit

The basic compressed coding unit directly records the readout of the sensor after exposure by linking readout electronics (with memory) to each pixel, as shown in Figure 2a.
In this paper, we propose to implement compressive sensing in a block-wise way, i.e., we design n × n ( n = 4 ) pixel block as the basic compressed coding unit and only use m ( m = 1 4 ) readout electronic (with memory) to record the 0−1 weighted sums of intensity distributions from n × n pixels. In other words, the value of each readout electronics (with memory) is the sum of the random pixels in the n × n area. This operation can be abstracted into the basic formula of compressed sensing:
y = A x + w ; A R m × n 2 , x R n 2 × 1 , y R m ,
where A R m × n 2 is the compressive measurement matrix consisting of 0 and 1, which denotes the designed circuit connection state between the readout circuits and the pixels in each n × n pixel unit. The connection settings within each block of pixels are shown in Figure 2c–f; if the readout circuit is connected with the pixel, the corresponding value in the measurement matrix is 1 and the readout circuit will collect the output of this pixel. If not, it is 0 and the readout circuit does not record the readout of this pixel. For example, in the basic compressed coding unit of our CS-SPAD, the measurement matrix A is shown in Figure 3.
x R n 2 × 1 is the compressed sensing signal to be sampled and is also equivalent to the number of photons arriving at the pixel within an exposure time in a basic compressed coding unit. y R m is the sampled signal in compressed sensing in a basic compressed coding unit.

3.1.2. CS-SPAD

Figure 2b shows the overall layout of the CS-SPAD chip we designed. The entire pixel array of CS-SPAD consists of the basic compressed coding unit described above. For each pixel in the basic compressed coding unit, we use a single-photon avalanche diode as a solid-state photodetector to record the number of photons arriving at each pixel. The readout is the sum of the photon counts of all its connected pixels. For each 4 × 4 SPAD local block, four readout circuits are required to capture the compressively sampled data, i.e., the use of readout circuits is also reduced to 25%.

3.2. Information Processing Architecture Based on Convolution Neural Network

We propose CSSPAD-Net to realize multi-task processing upon the compressed measurements from our chip, realizing both image reconstruction and classification, as shown in Figure 4.

3.2.1. Reconstruction Branch

The proposed CS-SPAD sensor compressed the data by dividing the image, of which the pixel resolution is 32 × 32 , into a basic processing unit. In the basic processing unit, 4 × 4 pixels are compressed into 1 × 4 values. Based on the basic processing unit, a fully connected layer, for which the number of input channels is 4 and the number of output channels is 16, is used for the recovery of compressed data dimensions in a basic processing unit. All basic CS processing units in the same compressed data are upsampled by using this fully connected layer. Once the above steps have been completed, the initial reconstruction of the scene is completed through tilling the reconstructed block images from different CS pixel units. We further refine the initially reconstructed image by using convolutional layers to eliminate the block artifacts caused by the block processing of the CS-SPAD. For further utilization of the statistical prior of nature images, we utilize a similar structure through a global–local residual learning way. Global residual learning [37] could enforce the overall goal of the network to learn the residual details of the initially reconstructed image and largely improve the learning difficulty compared to directly learning the image itself [38], and we introduce the local residual learning in the residual dense block (RDB) to further helps the fusion of deep and shallow features in the network.

3.2.2. Classification Branch

To realize efficient perception with the proposed CS-SPAD sensor, we propose to realize classification besides reconstruction. In the classification branch, we propose to use four residual blocks as the main body of the classification branch and a linear layer for final classification for efficiency. As is well known, in the image reconstruction network, the shallow layers of the network contain more low-level details and the deep layer of the network contains more high-level features [39]. Bridge operations between the two branches are introduced to guarantee the information fusion among different tasks.

3.2.3. Implementation Details

We synchronously train the reconstruction branch and the classification branch in our CSSPAD-Net together. The loss function in the reconstruction branch is the mean square error (MSE) loss L MSE , and we set the learning rate to 0.0004. Given reconstructed image x rec and the ground truth image x with N pixels, the MSE loss can be calculated as:
L MSE = 1 N x rec x F 2 .
where | | F ( . ) | | is the Frobenius norm. In the classification branch, the loss function is the cross entropy (CE) loss and the learning rate is set to 0.1. Given predicted vector x pred = [ x 1 , x 2 , , x C ] with C classes and the ground truth class j, the CE loss can be calculated as:
L CE = log e x j pred i = 1 C e x i pred .
We train our CSSPAD-Net on Nvidia GeForce RTX 2080 for 200 epochs and the learning rate of both is reduced by half every 50 epochs. The batch size is set to 128. The Adam optimizer [40] is adopted with β 1 = 0.9 , β 2 = 0.999 and ϵ = 1 × 10 8 .

4. Experiments

We verify the CS-SPAD sensor in real scenes and use CSSPAD-Net to complete scene reconstruction and perceptual classification. Unlike commercial cameras that integrate focusing and control systems, the CS-SPAD we proposed is just a computational sensor with a photoelectric conversion function, which requires an additional focusing lens and control system to cooperate with the CS-SPAD sensor. In this section, we first introduce our CS-SPAD sensor chip and the optical system, after which we will demonstrate the effectiveness of the CS-SPAD sensor and the proposed CSSPAD-Net with experimental results on MNIST data [22].

4.1. Prototype CS-SPAD Sensor Chip and the Optical System

We designed a 32 × 32 CS-SPAD sensor chip to realize on-chip snapshot-type spatial compressive imaging, as shown in Figure 5a. The system mainly includes three parts: a 32 × 32 SPAD detector array, readout circuits, and address decoding circuits. The 32 × 32 SPAD detector circuit includes the SPAD detector, the corresponding gating quenching circuit, and the logic circuit required for compressive sensing. The readout circuit includes a pulse shaper, a transmission circuit, a 12-bit counter circuit, and a 12-bit latch circuit. The address decoding circuit uses two identical 8 × 16 two-stage decoding modules, which work synchronously and transmit data through two 12-bit IO data ports.
The chip system architecture is shown in Figure 6, consisting of three main parts: a 32 × 32 SPAD detector array, 256-row readout circuits, and an address decoding circuit. The 32 × 32 SPAD detector circuit includes the SPAD detector, the corresponding gating quenching circuit, and the logic circuits for multiple “wired AND” operations required by compressed sensing. The row readout circuit includes a shaping circuit, a transmission circuit for integrating time, a 12-bit counter circuit, and a 12 bit-buffer circuit. The address decoding circuit adopts two identical 8 × 16 two-stage decoding modules, which work synchronously and transmit data through two 12-bit IO data ports.
The chip’s internal structure and the pixel layout are shown in Figure 5b,c, respectively. We adopt the 0.18 μ m 1P6M CMOS technology, and the SPAD pixel size is 15 μ m. The array size is 32 × 32 and the bit depth of the counter is 12 bit. The calibrated dark count of the proposed CS-SPAD is 200 cps at room temperature, i.e., 300 K. The dead time of the sensor is about 20 ns. Our CS-SPAD works in avalanche mode with single-photon sensitivity and the quantum efficiency is about 15%. The performance summary of the CS-SPAD sensor chip is shown in Table 1.
Additionally, as shown in Table 2, we compared different SPAD-based imagers or imaging systems with our CS-SPAD. We propose the on-chip compressed sensing method that performs compressed sampling directly on the chip, which can effectively avoid the complexity of the optical system.
As shown in Figure 7, we build a prototype imaging system based on the proposed CS-SPAD sensor. From left to right are the target scene, the lens used for focusing, the CS-SPAD sensor, and the host computer for controlling the automatic execution of the system. The working pipeline of the CS-SPAD imaging is as follows: first, the host computer controls the display to display a pattern to be sampled on the monitor, then the CSSPAD is exposed, and finally the compressed data sampled by the CS-SPAD sensor are read out and the reconstruction and classification results can be implemented with the trained network on the host computer.

4.2. Experiment Results

4.2.1. Dataset and End-to-End Network Training

We evaluate the CS-SPAD sensor and the CSSPAD-Net on the MNIST [22] dataset, a dataset of handwritten digits for classification tasks. We use the full MNIST dataset with 60,000 training images and 10,000 test images for the validation of our entire system.

4.2.2. Simulation and CSSPAD Sampling

For quantitative evaluation, with the test dataset, we display the images on the screen, capture the compressed data with the CS-SPAD sensor, reconstruct the image with the CSSPAD-Net, and calculate the reconstruction metrics with the reconstructed image and the corresponding projected image. The reconstruction and the classification results are shown in Table 3, we use PSNR (peak signal-to-noise ratio) and SSIM (structural similarity index measure) [51] to evaluate the quality of the reconstruction. As shown, our method could realize high classification accuracy and preserve almost all the details with high SSIM metrics. For quantitative evaluation, we further show the reconstruction result of the reconstruction branch in CSSPAD-Net in Figure 8. In the figure, “CS-SPAD” indicates that the acquisition is completed by the CS-SPAD chip, and the reconstruction process is completed on the GPU. Moreover, “simulation” indicates that the acquisition process simulates the CS-SPAD acquisition process, and the reconstruction is also completed on the GPU. As is shown, the structural details of different digits are elegantly recovered with fine details.
We also analyzed the reconstruction quality and classification accuracy of different categories of digital numbers, as shown in Figure 9 Although the PSNR of different categories of the reconstructed digital numbers fluctuates a little, it has little effect on the classification results of the classification branch.

4.2.3. Real Handwritten Data Experiment

Furthermore, as shown in Figure 10, we handwrite more digits, not from MNIST, for verification of our entire system. We use the same model trained on the MNIST dataset to classify and reconstruct the real handwritten digits. The visual display of the reconstructed results are shown in Figure 11, as shown, our methods could realize elegant reconstruction with the compressed results captured by the CS-SPAD sensor.

5. Conclusions

In this paper, we propose CS-SPAD to realize on-chip spatial compressive imaging, which could reduce the required sensor area for readout electronics and the data throughput to 25%. CSSPAD-Net is further proposed to recover images from compressively sampled data and implement high-accuracy perceptual classification. We taped out the CS-SPAD array and built a prototype imaging system to demonstrate the effectiveness of the sensor. Quantitative and qualitative experiments on MNIST dataset [22] and real hand-written digits were conducted to demonstrate the effectiveness and efficiency of our proposed CS-SPAD imaging sensor. For future work, we plan to further improve the performance of the CS-SPAD imaging chip and extend the on-chip CS idea to 3D imaging. Specifically, since existing research on compressed sensing [52,53,54,55,56] have shown that the joint optimization of the reconstruction network and the compressive sensing matrix can greatly improve the imaging efficiency, we plan to introduce the end-to-end optimization of the sensing matrix and reconstruction algorithm to further improve the CS imaging efficiency of our CS-SPAD imager. Beyond that, since a large-format SPAD array with a time-to-digital converter (TDC) module is in high demand for 3D imaging, which encounters more severe challenges of high data bandwidth and large sensor area for peripheral circuits of TDC, we plan to further develop an on-chip CS-SPAD with a TDC module, based on the proposed CS-SPAD method, to realize efficient 3D CS detection.

Author Contributions

Conceptualization, X.H. and T.Y.; methodology, C.Q.; sensor design and fabrication, X.K. and C.M.; software, C.Q. and P.W.; validation, C.Q. and P.W.; formal analysis, C.Q. and P.W.; investigation, C.Q. and P.W.; resources, C.Q. and P.W.; data curation, C.Q., X.H. and P.W.; writing—original draft preparation, C.Q. and P.W.; writing—review and editing, X.H.; visualization, C.Q.; supervision, X.H.; project administration, X.H., T.Y. and F.Y.; funding acquisition, X.H. and T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by NSFC Projects 61971465, National Key Research and Development Program of China (2022YFA1207200), and Fundamental Research Funds 244 for the Central Universities, China (Grant No. 0210-14380184).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Mnist [22] data presented in this study are openly available and can be founed at http://yann.lecun.com/exdb/mnist/, accessed on 27 April 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shin, D.; Xu, F.; Venkatraman, D.; Lussana, R.; Villa, F.; Zappa, F.; Goyal, V.K.; Wong, F.N.; Shapiro, J.H. Photon-efficient imaging with a single-photon camera. Nat. Commun. 2016, 7, 12046. [Google Scholar] [CrossRef] [PubMed]
  2. Ingle, A.; Velten, A.; Gupta, M. High flux passive imaging with single-photon sensors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6760–6769. [Google Scholar]
  3. Liu, Y.; Gutierrez-Barragan, F.; Ingle, A.; Gupta, M.; Velten, A. Single-photon camera guided extreme dynamic range imaging. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022; pp. 1575–1585. [Google Scholar]
  4. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.; Tardós, J.D. Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  5. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  6. Zhu, X.; Wang, Y.; Dai, J.; Yuan, L.; Wei, Y. Flow-guided feature aggregation for video object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 408–417. [Google Scholar]
  7. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  8. Carreira, J.; Zisserman, A. Quo vadis, action recognition? In A new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 6299–6308. [Google Scholar]
  9. Bruschini, C.; Homulle, H.; Antolovic, I.M.; Burri, S.; Charbon, E. Single-photon avalanche diode imagers in biophotonics: Review and outlook. Light. Sci. Appl. 2019, 8, 87. [Google Scholar] [CrossRef]
  10. Morimoto, K.; Ardelean, A.; Wu, M.L.; Ulku, A.C.; Antolovic, I.M.; Bruschini, C.; Charbon, E. Megapixel time-gated SPAD image sensor for 2D and 3D imaging applications. Optica 2020, 7, 346–354. [Google Scholar] [CrossRef]
  11. Morimoto, K.; Iwata, J.; Shinohara, M.; Sekine, H.; Abdelghafar, A.; Tsuchiya, H.; Kuroda, Y.; Tojima, K.; Endo, W.; Maehashi, Y.; et al. 3.2 megapixel 3D-stacked charge focusing SPAD for low-light imaging and depth sensing. In Proceedings of the 2021 IEEE International Electron Devices Meeting (IEDM), San Francisco, CA, USA, 11–16 December 2021; pp. 20–22. [Google Scholar]
  12. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  13. Brunton, S.L.; Proctor, J.L.; Tu, J.H.; Kutz, J.N. Compressed sensing and dynamic mode decomposition. J. Comput. Dyn. 2015, 2, 165. [Google Scholar] [CrossRef]
  14. McMackin, L.; Herman, M.A.; Chatterjee, B.; Weldon, M. A high-resolution SWIR camera via compressed sensing. In Proceedings of the Infrared Technology and Applications XXXVIII, Baltimore, MD, USA, 23–27 April 2012; Volume 8353, p. 835303. [Google Scholar]
  15. Dadkhah, M.; Deen, M.J.; Shirani, S. Block-based CS in a CMOS image sensor. IEEE Sens. J. 2012, 14, 2897–2909. [Google Scholar] [CrossRef]
  16. Katic, N.; Kamal, M.H.; Kilic, M.; Schmid, A.; Vandergheynst, P.; Leblebici, Y. Power-efficient CMOS image acquisition system based on compressive sampling. In Proceedings of the 2013 IEEE 56th International Midwest Symposium on Circuits and Systems (MWSCAS), Columbus, OH, USA, 4–7 August 2013; pp. 1367–1370. [Google Scholar]
  17. Majidzadeh, V.; Jacques, L.; Schmid, A.; Vandergheynst, P.; Leblebici, Y. A (256 × 256) pixel 76.7 mW CMOS imager/compressor based on real-time in-pixel compressive sensing. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 30 May–2 June 2010; pp. 2956–2959. [Google Scholar]
  18. Oike, Y.; El Gamal, A. CMOS image sensor with per-column ΣΔ ADC and programmable compressed sensing. IEEE J. Solid-State Circuits 2012, 48, 318–328. [Google Scholar] [CrossRef]
  19. Leitner, S.; Wang, H.; Tragoudas, S. Design of scalable hardware-efficient compressive sensing image sensors. IEEE Sens. J. 2017, 18, 641–651. [Google Scholar] [CrossRef]
  20. Lee, H.; Kim, W.T.; Kim, J.; Chu, M.; Lee, B.G. A compressive sensing CMOS image sensor with partition sampling technique. IEEE Trans. Ind. Electron. 2020, 68, 8874–8884. [Google Scholar] [CrossRef]
  21. Park, C.; Zhao, W.; Park, I.; Sun, N.; Chae, Y. A 51-pJ/pixel 33.7-dB PSNR 4× compressive CMOS image sensor with column-parallel single-shot compressive sensing. IEEE J. Solid-State Circuits 2021, 56, 2503–2515. [Google Scholar] [CrossRef]
  22. Deng, L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  23. Wayne, M.; Ulku, A.; Ardelean, A.; Mos, P.; Bruschini, C.; Charbon, E. A 500 × 500 dual-gate SPAD imager with 100% temporal aperture and 1 ns minimum gate length for FLIM and phasor imaging applications. IEEE Trans. Electron Devices 2022, 69, 2865–2872. [Google Scholar] [CrossRef]
  24. Bruschini, C.; Burri, S.; Bernasconi, E.; Milanese, T.; Ulku, A.C.; Homulle, H.; Charbon, E. LinoSPAD2: A 512x1 linear SPAD camera with system-level 135-ps SPTR and a reconfigurable computational engine for time-resolved single-photon imaging. In Proceedings of the Quantum Sensing and Nano Electronics and Photonics XIX, SPIE, San Francisco, CA, USA, 29 January–2 February 2023; Volume 12430, pp. 126–135. [Google Scholar]
  25. Ma, J.; Zhang, D.; Elgendy, O.A.; Masoodian, S. A 0.19 e-rms read noise 16.7 Mpixel stacked quanta image sensor with 1.1 μm-pitch backside illuminated pixels. IEEE Electron Device Lett. 2021, 42, 891–894. [Google Scholar] [CrossRef]
  26. Sundar, V.; Ma, S.; Sankaranarayanan, A.C.; Gupta, M. Single-Photon Structured Light. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2022; pp. 17865–17875. [Google Scholar]
  27. Gutierrez-Barragan, F.; Ingle, A.; Seets, T.; Gupta, M.; Velten, A. Compressive single-photon 3D cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2022; pp. 17854–17864. [Google Scholar]
  28. Hutchings, S.W.; Johnston, N.; Gyongy, I.; Al Abbas, T.; Dutton, N.A.; Tyler, M.; Chan, S.; Leach, J.; Henderson, R.K. A reconfigurable 3-D-stacked SPAD imager with in-pixel histogramming for flash LIDAR or high-speed time-of-flight imaging. IEEE J. -Solid-State Circuits 2019, 54, 2947–2956. [Google Scholar] [CrossRef]
  29. Candès, E.J.; Wakin, M.B. An introduction to compressive sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  30. Hu, X.; Suo, J.; Yue, T.; Bian, L.; Dai, Q. Patch-primitive driven compressive ghost imaging. Opt. Express 2015, 23, 11092–11104. [Google Scholar] [CrossRef] [PubMed]
  31. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 449–458. [Google Scholar]
  32. Katkovnik, V.; Astola, J. Compressive sensing computational ghost imaging. JOSA A 2012, 29, 1556–1567. [Google Scholar] [CrossRef]
  33. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef]
  34. Morris, P.A.; Aspden, R.S.; Bell, J.E.; Boyd, R.W.; Padgett, M.J. Imaging with a small number of photons. Nat. Commun. 2015, 6, 5913. [Google Scholar] [CrossRef] [PubMed]
  35. Wang, Y.; Huang, K.; Fang, J.; Yan, M.; Wu, E.; Zeng, H. Mid-infrared single-pixel imaging at the single-photon level. Nat. Commun. 2023, 14, 1073. [Google Scholar] [CrossRef]
  36. Sun, Q.; Dun, X.; Peng, Y.; Heidrich, W. Depth and transient imaging with compressive spad array cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 273–282. [Google Scholar]
  37. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  41. Parmesan, L.; Dutton, N.; Calder, N.J.; Krstajic, N.; Holmes, A.J.; Grant, L.A.; Henderson, R.K. A 256 × 256 SPAD array with in-pixel time to amplitude conversion for fluorescence lifetime imaging microscopy. Memory 2015, 900, M5. [Google Scholar]
  42. Perenzoni, M.; Massari, N.; Perenzoni, D.; Gasparini, L.; Stoppa, D. A 160 × 120 Pixel Analog-Counting Single-Photon Imager With Time-Gating and Self-Referenced Column-Parallel A/D Conversion for Fluorescence Lifetime Imaging. IEEE J. Solid-State Circuits 2015, 51, 155–167. [Google Scholar]
  43. Lee, C.; Johnson, B.; Jung, T.; Molnar, A. A 72 × 60 angle-sensitive SPAD imaging array for lens-less FLIM. Sensors 2016, 16, 1422. [Google Scholar] [CrossRef]
  44. Gyongy, I.; Calder, N.; Davies, A.; Dutton, N.A.; Duncan, R.R.; Rickman, C.; Dalgarno, P.; Henderson, R.K. A 256 × 256, 100-kfps, 61% Fill-Factor SPAD Image Sensor for Time-Resolved Microscopy Applications. IEEE Trans. Electron Devices 2017, 65, 547–554. [Google Scholar] [CrossRef]
  45. Burri, S.; Bruschini, C.; Charbon, E. LinoSPAD: A compact linear SPAD camera system with 64 FPGA-based TDC modules for versatile 50 ps resolution time-resolved imaging. Instruments 2017, 1, 6. [Google Scholar] [CrossRef]
  46. Zhang, C.; Lindner, S.; Antolovic, I.M.; Wolf, M.; Charbon, E. A CMOS SPAD imager with collision detection and 128 dynamically reallocating TDCs for single-photon counting and 3D time-of-flight imaging. Sensors 2018, 18, 4016. [Google Scholar] [CrossRef] [PubMed]
  47. Ulku, A.C.; Bruschini, C.; Antolović, I.M.; Kuo, Y.; Ankri, R.; Weiss, S.; Michalet, X.; Charbon, E. A 512 × 512 SPAD image sensor with integrated gating for widefield FLIM. IEEE J. Sel. Top. Quantum Electron. 2018, 25, 6801212. [Google Scholar] [CrossRef]
  48. Henderson, R.K.; Johnston, N.; Hutchings, S.W.; Gyongy, I.; Al Abbas, T.; Dutton, N.; Tyler, M.; Chan, S.; Leach, J. 5.7 A 256 × 256 40 nm/90 nm CMOS 3D-stacked 120 dB dynamic-range reconfigurable time-resolved SPAD imager. In Proceedings of the 2019 IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 17–21 February 2019; pp. 106–108. [Google Scholar]
  49. Okino, T.; Yamada, S.; Sakata, Y.; Kasuga, S.; Takemoto, M.; Nose, Y.; Koshida, H.; Tamaru, M.; Sugiura, Y.; Saito, S.; et al. 5.2 A 1200 × 900 6 μm 450 fps Geiger-mode vertical avalanche photodiodes CMOS image sensor for a 250 m time-of-flight ranging system using direct-indirect-mixed frame synthesis with configurable-depth-resolution down to 10 cm. In Proceedings of the 2020 IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 16–20 February 2020; pp. 96–98. [Google Scholar]
  50. Kumagai, O.; Ohmachi, J.; Matsumura, M.; Yagi, S.; Tayu, K.; Amagawa, K.; Matsukawa, T.; Ozawa, O.; Hirono, D.; Shinozuka, Y.; et al. 7.3 A 189 × 600 back-illuminated stacked SPAD direct time-of-flight depth sensor for automotive LiDAR systems. In Proceedings of the 2021 IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 13–22 February 2021; Volume 64, pp. 110–112. [Google Scholar]
  51. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  52. Shi, W.; Jiang, F.; Zhang, S.; Zhao, D. Deep networks for compressed image sensing. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 877–882. [Google Scholar]
  53. Chen, Z.; Guo, W.; Feng, Y.; Li, Y.; Zhao, C.; Ren, Y.; Shao, L. Deep-learned regularization and proximal operator for image compressive sensing. IEEE Trans. Image Process. 2021, 30, 7112–7126. [Google Scholar] [CrossRef]
  54. Cui, W.; Liu, S.; Jiang, F.; Zhao, D. Image Compressed Sensing Using Non-local Neural Network. IEEE Trans. Multimed. 2021, 25, 816–8303. [Google Scholar] [CrossRef]
  55. Fan, Z.E.; Lian, F.; Quan, J.N. Global Sensing and Measurements Reuse for Image Compressed Sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8954–8963. [Google Scholar]
  56. Mou, C.; Zhang, J. TransCL: Transformer makes strong and flexible compressive learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5236–5251. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Pipeline of the proposed CS-SPAD sensing method.
Figure 1. Pipeline of the proposed CS-SPAD sensing method.
Sensors 23 04417 g001
Figure 2. (a) Classic sensor array (b), basic compressed sensing imaging unit of CS-SPAD, (cf) four different CS connection settings.
Figure 2. (a) Classic sensor array (b), basic compressed sensing imaging unit of CS-SPAD, (cf) four different CS connection settings.
Sensors 23 04417 g002
Figure 3. An example of a CS-SPAD measurement matrix.
Figure 3. An example of a CS-SPAD measurement matrix.
Sensors 23 04417 g003
Figure 4. (a) Overview of CSSPAD-Net, (b) the structure of residual block (Resblock), (c) the structure of residual dense block (RDB).
Figure 4. (a) Overview of CSSPAD-Net, (b) the structure of residual block (Resblock), (c) the structure of residual dense block (RDB).
Sensors 23 04417 g004
Figure 5. (a) CS SPAD chip overview, (b) the diagram of CS-SPAD sensor chip design, (c) single pixel layout.
Figure 5. (a) CS SPAD chip overview, (b) the diagram of CS-SPAD sensor chip design, (c) single pixel layout.
Sensors 23 04417 g005
Figure 6. The CS SPAD chip architecture.
Figure 6. The CS SPAD chip architecture.
Sensors 23 04417 g006
Figure 7. (a) Overview of the prototype imaging system, (b) the detail of camera lens and CS SPAD chip, (c) the detail of peripheral circuits.
Figure 7. (a) Overview of the prototype imaging system, (b) the detail of camera lens and CS SPAD chip, (c) the detail of peripheral circuits.
Sensors 23 04417 g007
Figure 8. Reconstructed results by CSSPAD-Net.
Figure 8. Reconstructed results by CSSPAD-Net.
Sensors 23 04417 g008
Figure 9. Reconstruction and classification results for different digital number categories by CSSPAD-Net. (a) Reconstruction PSNR with different digital number categories and (b) classification accuracy with different digital number categories.
Figure 9. Reconstruction and classification results for different digital number categories by CSSPAD-Net. (a) Reconstruction PSNR with different digital number categories and (b) classification accuracy with different digital number categories.
Sensors 23 04417 g009
Figure 10. Example of real handwritten digits.
Figure 10. Example of real handwritten digits.
Sensors 23 04417 g010
Figure 11. Real handwritten digits: reconstructed results.
Figure 11. Real handwritten digits: reconstructed results.
Sensors 23 04417 g011
Table 1. Performance summary of the 32 × 32 CS-SPAD sensor chip.
Table 1. Performance summary of the 32 × 32 CS-SPAD sensor chip.
ParameterValue
Technology0.18 μ m 1P6M CMOS
Chip size2.9 mm × 3.3 mm
Array size32 × 32
Pixel size15 μ m
Counter width12 bit
Dark counts rate200 cps
Dead time20 ns
Power supply1.8 V (Digital)/3.3 V (Analog)
/12 V (SPAD cathode voltage)
Power consumption10 mW@12 V SPAD cathode voltage
Table 2. Overview of recently developed SPAD imagers or imaging systems.
Table 2. Overview of recently developed SPAD imagers or imaging systems.
SPAD Sensor or
Imaging System
YearCSCS
Methods
SPAD
Array
Technology
(nm)
Pixel Size
( μ m)
Dark Count Rate
(cps)
[41]2015NoNone 256 × 256 1304.230.8
[42]2016NoNone 160 × 160 35015580
[43]2016NoNone 72 × 60 180152.3
[1]2016YesOptical32 × 32350150100
[44]2017NoNone 256 × 256 13014.16200
[45]2017NoNone 256 × 1 35017.11286.6
[46]2018NoNone 32 × 32 18017113
[36]2018YesOptical 64 × 32 /30150
[47]2018NoNone 512 × 512 18067.5
[48]2019NoNone 256 × 256 40/909.2/38.420
[10]2020NoNone 1024 × 1000 1809.40.4/2.0
[49]2020NoNone 1200 × 900 656100
[11]2021NoNone 2072 × 1548 40/906.391.8
[50]2021NoNone 189 × 600 40/90102000
[27]2022YesSimulation////
[23]2022NoNone 500 × 500 18016.3810.2
[35]2023YesOptical 1 × 1 /180100
[24]2023NoNone 512 × 1 18026.2<100
Ours2023YesOn-chip32 × 3218015200
Table 3. Accuracy, PSNR, and SSIM of two datasets.
Table 3. Accuracy, PSNR, and SSIM of two datasets.
DatasetAverage PSNR/dBAverage SSIMAccuracy/%
CSSPAD27.30390.981999.22
Simulation31.67600.993099.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qiu, C.; Wang, P.; Kong, X.; Yan, F.; Mao, C.; Yue, T.; Hu, X. On-Chip Compressive Sensing with a Single-Photon Avalanche Diode Array. Sensors 2023, 23, 4417. https://doi.org/10.3390/s23094417

AMA Style

Qiu C, Wang P, Kong X, Yan F, Mao C, Yue T, Hu X. On-Chip Compressive Sensing with a Single-Photon Avalanche Diode Array. Sensors. 2023; 23(9):4417. https://doi.org/10.3390/s23094417

Chicago/Turabian Style

Qiu, Chenxi, Peng Wang, Xiangshun Kong, Feng Yan, Cheng Mao, Tao Yue, and Xuemei Hu. 2023. "On-Chip Compressive Sensing with a Single-Photon Avalanche Diode Array" Sensors 23, no. 9: 4417. https://doi.org/10.3390/s23094417

APA Style

Qiu, C., Wang, P., Kong, X., Yan, F., Mao, C., Yue, T., & Hu, X. (2023). On-Chip Compressive Sensing with a Single-Photon Avalanche Diode Array. Sensors, 23(9), 4417. https://doi.org/10.3390/s23094417

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop