Next Article in Journal
Coarse–Fine Registration of Point Cloud Based on New Improved Whale Optimization Algorithm and Iterative Closest Point Algorithm
Previous Article in Journal
The Crystal Chemistry of Voltaite-Group Minerals from Post-Volcanic and Anthropogenic Occurrences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Color Sensing and Image Reconstruction Using Intelligent Machine Learning Algorithm with PINIP Radial Junction Imager

1
School of Electronics Science and Engineering, Nanjing University, Nanjing 210023, China
2
College of Physical Science and Technology, Yangzhou University, Yangzhou 225002, China
*
Authors to whom correspondence should be addressed.
Symmetry 2023, 15(12), 2127; https://doi.org/10.3390/sym15122127
Submission received: 27 October 2023 / Revised: 26 November 2023 / Accepted: 28 November 2023 / Published: 29 November 2023

Abstract

:
The development of a filterless imager has been eagerly awaited to overcome the diffraction limit when pixel sizes decrease to subwavelength scales. We propose an architecture for a filterless imager based on a symmetric inversely stacked radial junction (RJ) PINIP photodetector over silicon nanowires (SiNWs), whereby the diameter of which is less than 500 nm, which preliminarily displays the capability of bias-selected and tunable spectrum responses to the R, G, and B color bands. Assisted via suitably trained deep learning algorithms, the imager can provide more accurate color discrimination and imaging capabilities. Here, we used KNN (k-nearest neighbor) and convolution neural network (CNN) methods to retrieve the RGB ratios from the measured photocurrent value based on the pre-trained bias-tuned spectrum responses and reconstructed the images with high accuracy. Further, we demonstrated the capability of restoring sub-sampling pictures via CNN with a U-net architecture, and satisfactory reconstruction was obtained even with a sampling ratio as low as 20%. Our imaging scheme cannot only be used for high-resolution imaging but can also pave the way for application in single-pixel imaging and compressive sensing.

1. Introduction

Image sensing is of great importance in our modern information age. It is widely used in a variety of electronic devices for surveillance and monitoring, machine vision, and for medical and biological testing. An image sensor is one of the key components of the digital imaging system, usually consisting of an image sensor, optical system, and digital processors [1,2]. Nowadays, image sensors usually comprise an array of pixels on silicon-based charge-coupled devices (CCDs) or complementary metal-oxide semiconductors (CMOSs), which convert light to electrical signals. To obtain color images, an array of color filters are employed and deposited on top of the image sensor, typically a two dimensional symmetric mosaic Beyer filter array with the unit pattern of RGGB (two Gs for green, one R for red, and one B for blue) [3,4,5,6]. Dramatic progress has been achieved for this kind of pixelated image sensor. Conventional color image sensors already have record pixel densities above 100 megapixels per chip, where the color sensor pixel area features dimensions of about 2 × 2 μm2 [7]. However, it is very difficult to increase the pixel density for conventional pixelated color image sensors in order to achieve finer resolution imaging due to the diffraction limit. Moreover, the traditional approach of one pixel with four filters largely reduces the incoming light flux, resulting in low sensitivity and limiting further miniaturization to shrink the unit size. To break this limit, novel imaging schemes or even filterless imaging architectures have been proposed and demonstrated over the past decade [8]. For example, in metal nano-disk imagers, metal nano-disks are used to confine light on the nanoscale through the plasmon effect and to distinguish color according to the dispersion of plasmon resonance peaks related to the size of the nano-disk [9,10]. Silicon nanowire PN junction imagers consist of vertical silicon nanowire PN junction arrays with different diameters, which cannot only discriminate color, but also convert light into electrical signals by themselves. The color is detected/selected through the size-dependent cavity mode resonance effect [11,12,13,14]. There are also semiconductor nanoparticle imagers [15], where colloidal quantum dots (CQDs) of different sizes act as multiband filters because of the quantum confinement effect.
Besides the conventional imaging scheme, computational imaging (CI) using algorithms empowers the versatilities of imaging, which began with the restoration of imperfect images in medical diagnosis. With the aid of the machine deep learning (DL) approach, high-quality image restoration has been achieved in terms of accuracy, high resolution, and fast restoration speed [16]. DL is currently widely used in many fields and research areas, especially for use in image or spectral recognition tools [17]. First, the DL approach has the ability to exploit information from data that may be indiscernible using traditional methods. Second, its flexibility in design makes it compatible with nanophotonic platforms. Third, DL algorithms can be applicable to various functions, such as spectral reconstruction, high-resolution imaging, classification, and noise suppression. DL also boosts single-pixel imaging (SPI) where signals are collected by a single detector in a raster scanning way or by interrogating the scene using a series of patterns (the Hadamard pattern, for example); based on these data/patterns, the original image is reconstructed via the machine DL algorithm [18].
In recent work, we demonstrated a filterless imager based on a symmetric inversely stacked radial junction (RJ) PINIP (here, P, I, and N stand for p-type, intrinsic, and n-type amorphous Si film) photodetector fabricated around silicon nanowires (SiNWs) [19], where different incident wavelengths are absorbed with different penetration depths in stacked PIN junctions (as shown in the inset of Figure 1a). Such a 3D structure can confine light within the nanowire and redistribute the light field into different absorber layers [20], where the inner layer can absorb long wavelengths while the outer layer can absorb short wavelengths. Thus, this RJ PINIP sensor has a preliminary capability of bias-selected and tunable spectrum responses to the R, G, and B color bands, and it may be the smallest multiplex color sensor on the nanoscale [21,22]. By distinguishing the dispersion of the photocurrent under different bias voltages, this imager can deduce the RGB ratio from the measured photocurrent values. However, different RGB combinations or light with different power can give the same photocurrent value; thus, how to precisely discriminate the RGB color based on photocurrent values under different external bias voltages is the key problem. To address these issues, we employ the DL approach, which is widely used in image restoration and in the miniaturization of modern technologies [23,24,25,26].
Therefore, in this work, we use DL information process models such as k-nearest neighbors (KNNs) and convolutional neural networks (CNNs) to determine the RGB ratios from the measured photocurrent value based on the pre-trained bias-tuned spectrum responses, as well as conduct image restoration with higher accuracy. Further, using CNN to restore under-sampled pictures, a satisfactory image is reconstructed with only 20% sampling. These results highlight that our imaging paradigm combining unique RJ PINIP architecture and the DL approach can efficiently discriminate color and construct color images with high accuracy, laying the building blocks using an RJ PINIP unit with a diameter of around 500 nm for highly dense pixel color imaging as well as paving the way for applications in single-pixel imaging and compressive sensing.
The remainder of this paper is organized as follows. Section 2 contains the main body of our work, first introducing the architecture and bias-tuning color discrimination mechanism of our imager; then, using our imager aided by KNN and CNN methods, we demonstrate how to carry out the color recognition and image restoration, as well as make a comparison between the two methods. Finally, we restore the under-sampled image using the CNN. Section 3 gives a brief discussion and summary of the paper.

2. Experiments and Results

Figure 1a shows the scanning electron microscopy (SEM) image of the final fabricated RJ PINIP imagers, measuring ~500 nm in diameter and 1 μ m in length. The inset schematically draws the radially stacked PINIP multilayer structure. The simplified fabrication process is briefly described as follows. First, we fabricated the center boron (B)-doped p-type SiNWs upon (aluminum-doped ZnO) AZO glass substrates via a vapor–liquid–solid (VLS) method using the low-melting-point metal of tin (Sn) as a catalyst in a plasma-enhanced chemical vapor deposition (PECVD) system. As AZO is conductive, it acted as the bottom electrode. Next, the other layers of I/N/I/P were used for the a-Si:H layers with corresponding doping types; the inner and the outer intrinsic layers were both 50 nm and the n-type layer was 5 nm. Lastly, a 50 nm transparent and conductive ITO layer was coated by shadow masks in a magnetron sputtering system to define the top electrodes. Fabrication details can be found in Reference [13]. Figure 1b presents the external quantum efficiency (EQE) responses under different bias voltages [27]. It was found that the photocurrent response spectra could be continuously tuned by the bias voltage, indicating that the absorption center peak of the wavelength could be continuously modulated by the external bias (note that positive and negative photocurrents just mean opposite current directions). The photocurrent has a dispersion with the external bias voltage mainly due to the bias-selected absorption layer, with the inner or outer layer corresponding to the long or short wavelength, whereby each of which has unique responsivity with rich color information. This is the key idea for the RJ PINIP color discrimination imager.

2.1. Bias-Tunable Color Response of the Imager

We pre-trained our imager to determine the relationship between the photocurrent value under different biases and the known RGB ratios. Figure 2a shows the pre-trained setup. We used a computer-controlled projector to supply light with different RGB ratios. The RJ imager is placed in a fixed position in front of the projector, the bias is supplied by a source meter, and the data are collected automatically by the computer. Figure 2b shows the imager’s photocurrent response curves, which can be interpreted as the photocurrent value vs. bias voltage for a given color at specific RGB ratios. There is a slight distortion of the curves due to unknown vibrations in the environment. For each color from 0 to 255, increasing by 51 steps, we used 125 combinations of the RGB color. For each combination, we chose a bias voltage from 0 to 0.5 V in steps of 0.1 V and recorded the photocurrent value accordingly. We examined the photocurrent distribution and correlation, as shown in Figure 3. The diagonal spike pictures show the distribution of the current data at several voltages. The abscissa is the current. The ordinate is the number. The higher the number is, the more times this current appears. It can be seen that the current distribution pattern is different under different voltages. The other graphs of the scatter plot in Figure 3 outside the diagonal line are the photocurrent data under different voltages (the abscissa voltage A and the ordinate voltage B indicate that the scatter points on this graph are based on the photocurrent values under voltage A, as well as the abscissa and voltage B). The color of each point is the real color of the light illuminating the sample. It can be seen that these scattered points are relatively dispersive and not all concentrated on the diagonal. This means that these currents are not linearly related and can be used as independent, characteristic data for training.

2.2. Color Recognition and Image Restoration via KNN Method

The k-nearest neighbor (KNN) algorithm is an important tool in the field of machine learning [28,29]. The principle of KNN is to extract each specific parameter from an n-dimensional dataset and map it to an n-dimensional dataset, thereby achieving effective analysis of the data. When predicting a new value x, it determines which category x belongs to based on the categories of the k-nearest points, as illustrated in Figure 4. In addition, KNN can also be used for model optimization and data reconstruction to improve the accuracy and reliability of the model.
A photo of ‘Lena’ was pixelated and divided into 128 × 128 pixels for simplification (see Figure 5a). The pixel RGB color information was automatically scanned and projected onto the RJ-PINIP imager, as shown in Figure 2a, where the photocurrent signals Ibias were collected under different bias voltages. Then, these photocurrent signals were put together as a vector of Ivec, = (I0, I0.1, I0.2, I0.3, ……), where the subscript is the bias voltage and the number of Ivec depends on the number of biases used. First, we adopted the KNN classification method to guess the color (the R, G, and B values) from each value of vector Ivec with K = 2, based on the pre-trained knowledge shown in Figure 3 where photocurrents of each known RGB combination under different biased voltages are given. The RGB value of every pixel was obtained and, thus, the image was reconstructed, as shown in Figure 5b–d, with the photocurrent under 0–0.1 V, 0–0.2 V, and 0–0.3 V, respectively, and the step = 0.1 V. We can see that the image can be reconstructed with only two sets of photocurrent values between 0 and 0.1 V, but obviously, there is some false color which comes from sub-sampling, resulting in a larger L2 distance [30]; as a matter of fact, the more sets of photocurrent values, the better the restoration, but if the bias is above 0.4 V, there will be more false color due to the fact that there will be too many of the same photocurrent values for one RGB combination under different types of bias. The K value also has a great influence on the restoration effect [31]. We used the loss function to evaluate the recovery accuracy as a percentage, and as shown in Figure 5e, four sets of photocurrents and K = 2 form the optimal solution. In fact, as shown in Figure 3, most of the photocurrents under 0.4 V and 0.5 V are linear with the other photocurrents under different biases, indicating that more similar photocurrents occur, which makes it difficult for the KNN methods to obtain an accurate RGB value. A proper photocurrent number and proper K value can result in better image restoration. It is worth noting that the restored image is purplish, mainly due to the color tablet of the projector being rich in purple.

2.3. Color Recognition and Image Restoration Using CNN Method

We used the KNN algorithm to carry out the color recognition and restore the image, and the restored image is closer to the original. However, there are still many false colors and the restored image is not as smooth as it could be. In addition, this method is highly dependent on the training dataset, as the categories distinguished by the KNN algorithm are limited to existing pixel values in the training set [32]. In the field of image recognition and target detection, the CNN algorithm performs well [24]. It can learn hierarchical features, including high-level features, and is a key mechanism in feature extraction. For the CNN model, taking the current value corresponding to the input voltage of 0 to 0.3 V as an example, as sketched in Figure 6a, the corresponding input features are 0, I1, 0.1, I2, 0.2, I3, 0.3, and I4, and the number of channels of the input data is eight. The output features are R, G, and B, and the number of output channels is three. Since the input feature and output feature structures of the training data are relatively simple, in order to reduce the computational cost, we adopted a relatively simple CNN model structure of two convolutional layers and two fully connected layers. First, the first convolutional layer uses eight input feature channels and 12 filters, and performs convolution operations with three continuous input features (convolution kernel size). The stride is one, the padding is one, and the output size is 12 feature channels. Then, the output is processed nonlinearly through the ReLU (Rectified Linear Unit) activation function layer, which can set negative values to zero. Next, the second convolutional layer receives the output of the previous layer, with twelve input feature channels and six filters, using three consecutive input features for convolution operations. The stride is one, the padding is one, and the output size is six feature channels. Subsequently, the first fully connected layer maps these six features to a 256-dimensional feature space, performs linear transformation, and processes the output through the ReLU activation function. Finally, the second fully connected layer maps the 256-dimensional features to a 3-dimensional (corresponding to three RGB eigenvalues) feature space, performs linear transformation, and processes the output through the ReLU activation function. A simplified module is illustrated in Figure 6b.
The reconstructed images are given in Figure 7a–e using two to six sets of photocurrent data. The loss value of these two methods is given in Table 1 according to Formula (1).
Loss = i = 1 n expected   RGB   Value real   RGB   value 2 n
These images restored using the CNN are obviously better than those restored using the KNN method, due to the algorithm’s success in feature extraction and the activation function layer for rectification [33,34].

2.4. Sub-Sampling Image Restoration Using CNN Method

Our RJ-PINIP unit could be used as a single-pixel imager [35] and applied in compressive sensing [36]. In these scenarios, the raster scan strategy of sub-sampling is an efficient means of imaging, supported by a deep machine learning image reconstruction algorithm, resulting in a fast and super-resolution image. Using the CNN algorithm, we attempted to reconstruct Lena’s visage using 70% to 10% pixels. We built our CNN network in a conventional U-net structure based on the Keras deep learning model [37]. The U-net structure is a classic convolutional neural network structure [38] consisting of a set of symmetrical encoders (down-sampling paths) and decoders (up-sampling paths), connected through skip connections. The main advantage of this structure is that it can handle local and global features of images, and the introduction of skip connections helps to better recover details and mitigate information loss. This structure has been widely used and achieved good results in the fields of medical image segmentation [39], semantic segmentation [40,41], and image reconstruction [42]. The encoder part consists of multiple convolutional layers and pooling layers, which gradually extract the features of the image and reduce the size of the feature map. In the decoder part, the size of the feature map is restored to the size of the input image through an up-sampling operation, and skip connections are used to connect the output of the encoder with the input of the decoder to fuse feature information of different scales. The U-net structure designed here contains five down-sampling stages and four up-sampling stages. Each stage contains two convolutional layers, and the number of output channels of the last up-sampling stage is three, that is, the image is divided into RGB channels. Finally, the Adam optimizer is used to optimize the model, and the mean squared error (MSE) function is used as the objective function of training. We can observe the loss function to obtain the optimized image. First, we pre-trained this model with hundreds of known perfect white female face images and their corresponding damaged images (with 30–90% pixel loss, respectively). All of the original images had the same size of 128 × 128 pixels. Then, we obtained Lena’s distorted images by setting 40–90% of the pixel values to zero. To optimize the accuracy of the restoration, we first carried out the iteration test on an image with 30% pixel loss, as given in Figure 8; when the epoch = 100, the image is almost the same as its original with a small-enough MSE value.
Following this, we restored the damaged images to different degrees (40–90% pixel loss) with epoch = 100, and the results are shown in Figure 9. The images can be restored even with 80% pixel loss, and when the pixel loss is 90%, there are distorted features in the restored image, as well as false color. These results indicate that the PINIP radial junction imager only needs to randomly collect 20% of the pixel information to restore the overall facial structure. Our tiny RJ PINIP imager is expected to overcome the diffraction limit to realize super-fine imaging, and can also be used as a single-pixel imager application for holography [1], phase imaging [42], ophthalmic imaging [43], and ultrafast imaging [18]. This algorithm could greatly improve the single-pixel imaging efficiency of the PINIP radial junction imager.

3. Conclusions

Beyer-filter-based color imaging pixels reach a critical size in terms of subwavelength. In order to break the diffraction limit, we proposed a novel imaging scheme as well as the smallest color imaging unit in the world using an inversely stacked PINIP radial junction, whereby the spectrum response of which can be continuously tuned via the bias voltage. The responsivity under each bias is unique and rich in features, laying the basis for image reconstruction using the proper algorithm. Supported by deep machine learning algorithms, this imager cannot only distinguish color but can also sense and reconstruct the color image. In the KNN color classification method, the proper bias (0–0.3 V) and K value (K = 2) help to achieve higher color retrieval accuracy, while the CNN algorithm achieves better color recognition and image restoration. A satisfactory color image was reconstructed using only 20% of the pixel information, laying the foundation for single-pixel imaging and compressive sensing.

Author Contributions

J.W. conceived the project. S.Z. and Y.Z. fabricated and measured the samples. J.W., S.Z., and Y.Z. analyzed the data. Z.C. and Y.Z. performed the modeling. J.W. and Y.Z. wrote the manuscript. J.W. and S.Z. supervised the research. All authors discussed the obtained results. All authors have read and agreed to the published version of the manuscript.

Funding

This project was supported by the National Natural Science Foundation of China (grant nos. 61974064 and 61934004) and the National Key Research Program of China (grant no. 92164201).

Data Availability Statement

The datasets and code can be made available upon request to the corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Clemente, P.; Durán, V.; Tajahuerce, E.; Andrés, P.; Climent, V.; Lancis, J. Compressive holography with a single-pixel detector. Opt. Lett. 2013, 38, 2524–2527. [Google Scholar] [CrossRef] [PubMed]
  2. Sukhavasi, S.B.; Sukhavasi, S.B.; Elleithy, K.; Abuzneid, S.; Elleithy, A. CMOS Image Sensors in Surveillance System Applications. Sensors 2021, 21, 488. [Google Scholar] [CrossRef] [PubMed]
  3. Fossum, E.R. CMOS image sensors: Electronic camera-on-a-chip. IEEE Trans. Electron Devices 1997, 44, 1689–1698. [Google Scholar] [CrossRef]
  4. Huo, Y.; Fesenmaier, C.C.; Catrysse, P.B. Microlens performance limits in sub-2μm pixel CMOS image sensors. Opt. Express 2010, 18, 5861–5872. [Google Scholar] [CrossRef] [PubMed]
  5. Lindstrand, M. Sensor interpixel correlation analysis and reduction for color filter array high dynamic range image reconstruction. Color Res. Appl. 2019, 44, 335–347. [Google Scholar] [CrossRef]
  6. Goossens, S.; Navickaite, G.; Monasterio, C.; Gupta, S.; Piqueras, J.J.; Pérez, R.; Burwell, G.; Nikitskiy, I.; Lasanta, T.; Galán, T.; et al. Broadband image sensor array based on graphene–CMOS integration. Nat. Photonics 2017, 11, 366–371. [Google Scholar] [CrossRef]
  7. Lee, J.; Park, Y.; Kim, H.; Yoon, Y.-Z.; Ko, W.; Bae, K.; Lee, J.-Y.; Choo, H.; Roh, Y.-G. Compact meta-spectral image sensor for mobile applications. Nanophotonics 2022, 11, 2563–2569. [Google Scholar] [CrossRef]
  8. Wang, Z.; Yi, S.; Chen, A.; Zhou, M.; Luk, T.S.; James, A.; Nogan, J.; Ross, W.; Joe, G.; Shahsafi, A.; et al. Single-shot on-chip spectral sensors based on photonic crystal slabs. Nat. Commun. 2019, 10, 1020. [Google Scholar] [CrossRef]
  9. Kumar, K.; Duan, H.; Hegde, R.S.; Koh, S.C.W.; Wei, J.N.; Yang, J.K.W. Printing colour at the optical diffraction limit. Nat. Nanotechnol. 2012, 7, 557–561. [Google Scholar] [CrossRef]
  10. Tua, D.; Liu, R.; Yang, W.; Zhou, L.; Song, H.; Ying, L.; Gan, Q. Imaging-based intelligent spectrometer on a plasmonic rainbow chip. Nat. Commun. 2023, 14, 1902. [Google Scholar] [CrossRef]
  11. Tian, B.; Zheng, X.; Kempa, T.J.; Fang, Y.; Yu, N.; Yu, G.; Huang, J.; Lieber, C.M. Coaxial silicon nanowires as solar cells and nanoelectronic power sources. Nature 2007, 449, 885–889. [Google Scholar] [CrossRef] [PubMed]
  12. Park, H.; Dan, Y.; Seo, K.; Yu, Y.J.; Duane, P.K.; Wober, M.; Crozier, K.B. Filter-Free Image Sensor Pixels Comprising Silicon Nanowires with Selective Color Absorption. Nano Lett. 2014, 14, 1804–1809. [Google Scholar] [CrossRef] [PubMed]
  13. Meng, J.; Cadusch, J.J.; Crozier, K.B. Detector-Only Spectrometer Based on Structurally Colored Silicon Nanowires and a Reconstruction Algorithm. Nano Lett. 2020, 20, 320–328. [Google Scholar] [CrossRef] [PubMed]
  14. Cao, L.; Fan, P.; Barnard, E.S.; Brown, A.M.; Brongersma, M.L. Tuning the Color of Silicon Nanostructures. Nano Lett. 2010, 10, 2649–2654. [Google Scholar] [CrossRef] [PubMed]
  15. Bao, J.; Bawendi, M.G. A colloidal quantum dot spectrometer. Nature 2015, 523, 67–70. [Google Scholar] [CrossRef]
  16. Yousef, R.; Gupta, G.; Yousef, N.; Khari, M. A holistic overview of deep learning approach in medical imaging. Multimed. Syst. 2022, 28, 881–914. [Google Scholar] [CrossRef] [PubMed]
  17. Agarwal, S.; Jung, K.-H. Exploring Symmetry in Digital Image Forensics Using a Lightweight Deep-Learning Hybrid Model for Multiple Smoothing Operators. Symmetry 2023, 15, 2096. [Google Scholar] [CrossRef]
  18. Guo, Q.; Chen, H.; Weng, Z.; Chen, M.; Yang, S.; Xie, S. Compressive sensing based high-speed time-stretch optical microscopy for two-dimensional image acquisition. Opt. Express 2015, 23, 29639–29646. [Google Scholar] [CrossRef]
  19. Zhang, T.; Cao, L.; Zhang, S.; Liu, Z.; Wang, J.; Shi, Y.; Xu, J.; Chen, K.; Yu, L. Bias-selected full Red/Green/Blue color sensing and imaging based on inversely stacked radial PINIP junctions. Nano Futures 2020, 4, 035007. [Google Scholar] [CrossRef]
  20. Yang, F.; Wang, J.; Lu, J.; Yu, Z.; Yu, L.; Xu, J.; Shi, Y.; Chen, K.; Roca i Cabarrocas, P. Biomimetic Radial Tandem Junction Photodetector with Natural RGB Color Discrimination Capability. Adv. Opt. Mater. 2017, 5, 1700390. [Google Scholar] [CrossRef]
  21. Yang, Z.; Albrow-Owen, T.; Cai, W.; Hasan, T. Miniaturization of optical spectrometers. Science 2021, 371, 480. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, J.; Zheng, B.; Wang, X. Strategies for high performance and scalable on-chip spectrometers. J. Phys. Photonics 2021, 3, 012006. [Google Scholar] [CrossRef]
  23. Aghdam, H.H.; Heravi, E.J. Guide to Convolutional Neural Networks; Springer: New York, NY, USA, 2017; Volume 10, p. 51. [Google Scholar]
  24. Habibi Aghdam, H.; Jahani Heravi, E. (Eds.) Convolutional Neural Networks, in Guide to Convolutional Neural Networks: A Practical Application to Traffic-Sign Detection and Classification; Springer International Publishing: Cham, Switzerland, 2017; pp. 85–130. [Google Scholar]
  25. Zhou, Y.T.; Chellappa, R.; Vaid, A.; Jenkins, B.K. Image restoration using a neural network. IEEE Trans. Acoust. Speech Signal Process. 1988, 36, 1141–1151. [Google Scholar] [CrossRef]
  26. Ning, H.; Yu, Z.; Zhang, Q.; Wen, H.; Gao, B.; Mao, Y.; Li, Y.; Zhou, Y.; Zhou, Y.; Chen, J.; et al. An in-memory computing architecture based on a duplex two-dimensional material structure for in situ machine learning. Nat. Nanotechnol. 2023, 18, 493–500. [Google Scholar] [CrossRef] [PubMed]
  27. Ghosh, S.; Mukhopadhyay, B.; Sen, G.; Basu, P.K. Study of Si-Ge-Sn based Heterobipolar Phototransistor (HPT) exploiting Quantum Confined Stark Effect and Franz Keldysh effect with and without resonant cavity. Phys. E Low-Dimens. Syst. Nanostruct. 2019, 106, 62–67. [Google Scholar] [CrossRef]
  28. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Wang, R. Efficient kNN Classification with Different Numbers of Nearest Neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1774–1785. [Google Scholar] [CrossRef] [PubMed]
  29. Demidova, L.A. Two-Stage Hybrid Data Classifiers Based on SVM and kNN Algorithms. Symmetry 2021, 13, 615. [Google Scholar] [CrossRef]
  30. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN Model-Based Approach in Classification. In On the Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE; Springer: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  31. Liu, Q.; Liu, C. A Novel Locally Linear KNN Method with Applications to Visual Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2010–2021. [Google Scholar] [CrossRef]
  32. Hossain, E.; Hossain, M.F.; Rahaman, M.A. A Color and Texture Based Approach for the Detection and Classification of Plant Leaf Disease Using KNN Classifier. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’s Bazar, Bangladesh, 7–9 February 2019. [Google Scholar]
  33. Wahab, M.N.A.; Nazir, A.; Ren, A.T.Z.; Noor, M.H.M.; Akbar, M.F.; Mohamed, A.S.A. Efficientnet-Lite and Hybrid CNN-KNN Implementation for Facial Expression Recognition on Raspberry Pi. IEEE Access 2021, 9, 134065–134080. [Google Scholar] [CrossRef]
  34. Edgar, M.P.; Gibson, G.M.; Padgett, M.J. Principles and prospects for single-pixel imaging. Nat. Photonics 2019, 13, 13–20. [Google Scholar] [CrossRef]
  35. Starling, D.J.; Storer, I.; Howland, G.A. Compressive sensing spectroscopy with a single pixel camera. Appl. Opt. 2016, 55, 5198–5202. [Google Scholar] [CrossRef] [PubMed]
  36. Yin, X.-X.; Sun, L.; Fu, Y.; Lu, R.; Zhang, Y. U-Net-Based Medical Image Segmentation. J. Healthc. Eng. 2022, 2022, 4189781. [Google Scholar] [CrossRef] [PubMed]
  37. Konovalenko, I.; Maruschak, P.; Brezinová, J.; Prentkovskis, O.; Brezina, J. Research of U-Net-Based CNN Architectures for Metal Surface Defect Detection. Machines 2022, 10, 327. [Google Scholar] [CrossRef]
  38. Rahman, Z.; Zhang, R.; Bhutto, J.A. A Symmetrical Approach to Brain Tumor Segmentation in MRI Using Deep Learning and Threefold Attention Mechanism. Symmetry 2023, 15, 1912. [Google Scholar] [CrossRef]
  39. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar]
  40. Sambyal, N.; Saini, P.; Syal, R.; Gupta, V. Modified U-Net architecture for semantic segmentation of diabetic retinopathy images. Biocybern. Biomed. Eng. 2020, 40, 1094–1109. [Google Scholar] [CrossRef]
  41. Mahmoud, A.; Amin, K.; Al Rahhal, M.M.; Elkilani, W.S.; Mekhalfi, M.L.; Ibrahim, M. A CNN Approach for Emotion Recognition via EEG. Symmetry 2023, 15, 1822. [Google Scholar] [CrossRef]
  42. Soldevila, F.; Durán, V.; Clemente, P.; Lancis, J.; Tajahuerce, E. Phase imaging by spatial wavefront sampling. Optica 2018, 5, 164–174. [Google Scholar] [CrossRef]
  43. Lochocki, B.; Gambín, A.; Manzanera, S.; Irles, E.; Tajahuerce, E.; Lancis, J.; Artal, P. Single pixel camera ophthalmoscope. Optica 2016, 3, 1056–1059. [Google Scholar] [CrossRef]
Figure 1. (a) SEM photo of the RJ PINIP imager (inset: the schematic structure of the RJ PINIP unit) and (b) the EQE map of this imager under different biased voltages.
Figure 1. (a) SEM photo of the RJ PINIP imager (inset: the schematic structure of the RJ PINIP unit) and (b) the EQE map of this imager under different biased voltages.
Symmetry 15 02127 g001
Figure 2. (a) Measurement setup and (b) color-induced photocurrent vs. applied voltage to the imager illuminated via a projector.
Figure 2. (a) Measurement setup and (b) color-induced photocurrent vs. applied voltage to the imager illuminated via a projector.
Symmetry 15 02127 g002
Figure 3. Correlation between the photocurrents under different RGB ratios (shown as scattered colorful dots) with the biased voltages. The unit of the label of x and y is in μA. The spike diagrams along the diagonal are the statistical number of occurrences of the same photocurrent under different biased voltages.
Figure 3. Correlation between the photocurrents under different RGB ratios (shown as scattered colorful dots) with the biased voltages. The unit of the label of x and y is in μA. The spike diagrams along the diagonal are the statistical number of occurrences of the same photocurrent under different biased voltages.
Symmetry 15 02127 g003
Figure 4. Illustration of the principle of KNN classification with k = 3 scenario, square and triangle are two kinds of known different events, and star is unknown event.
Figure 4. Illustration of the principle of KNN classification with k = 3 scenario, square and triangle are two kinds of known different events, and star is unknown event.
Symmetry 15 02127 g004
Figure 5. (a) Original image and the reconstructed ones using KNN methods with data obtained under different biased voltages of (b) 0 and 0.1 V, (c) 0, 0.1 and 0.2 V, and (d) 0, 0.1, 0.2. and 0.3 V, and (e) 3D mapping of restoration accuracy with used bias number and K value.
Figure 5. (a) Original image and the reconstructed ones using KNN methods with data obtained under different biased voltages of (b) 0 and 0.1 V, (c) 0, 0.1 and 0.2 V, and (d) 0, 0.1, 0.2. and 0.3 V, and (e) 3D mapping of restoration accuracy with used bias number and K value.
Symmetry 15 02127 g005
Figure 6. (a) Network and (b) module of the CNN algorithm.
Figure 6. (a) Network and (b) module of the CNN algorithm.
Symmetry 15 02127 g006
Figure 7. Reconstructed images using CNN methods with data obtained under different biased voltages: (a) original, (b) 0–0.1 V, (c) 0–0.2 V, (d) 0–0.3 V, (e) 0–0.4 V, and (f) 0–0.5 V with a step of 0.1 V.
Figure 7. Reconstructed images using CNN methods with data obtained under different biased voltages: (a) original, (b) 0–0.1 V, (c) 0–0.2 V, (d) 0–0.3 V, (e) 0–0.4 V, and (f) 0–0.5 V with a step of 0.1 V.
Symmetry 15 02127 g007
Figure 8. (a)The original and (b) input of damaged image with 30% pixel loss and the output images restored using the CNN module with (c) epoch = 50 and (d) epoch = 100 and (e) the loss value curve with epoch.
Figure 8. (a)The original and (b) input of damaged image with 30% pixel loss and the output images restored using the CNN module with (c) epoch = 50 and (d) epoch = 100 and (e) the loss value curve with epoch.
Symmetry 15 02127 g008
Figure 9. Damaged images with different percentages of pixel loss from 40 to 90%, given as inputs (af), restored using the CNN module with epoch = 100, and shown as corresponding output.
Figure 9. Damaged images with different percentages of pixel loss from 40 to 90%, given as inputs (af), restored using the CNN module with epoch = 100, and shown as corresponding output.
Symmetry 15 02127 g009
Table 1. Comparisons of loss values between KNN and CNN algorithms regarding the restoration of the image with data obtained under different biases.
Table 1. Comparisons of loss values between KNN and CNN algorithms regarding the restoration of the image with data obtained under different biases.
Loss Value0~0.1 V0~0.2 V0~0.3 V0~0.4 V0~0.5 V
KNN30.1722.2124.5822.3222.67
CNN15.5813.4012.1312.7413.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Chen, Z.; Zhang, S.; Wang, J. Color Sensing and Image Reconstruction Using Intelligent Machine Learning Algorithm with PINIP Radial Junction Imager. Symmetry 2023, 15, 2127. https://doi.org/10.3390/sym15122127

AMA Style

Zhang Y, Chen Z, Zhang S, Wang J. Color Sensing and Image Reconstruction Using Intelligent Machine Learning Algorithm with PINIP Radial Junction Imager. Symmetry. 2023; 15(12):2127. https://doi.org/10.3390/sym15122127

Chicago/Turabian Style

Zhang, Yifei, Zongsen Chen, Shaobo Zhang, and Junzhuan Wang. 2023. "Color Sensing and Image Reconstruction Using Intelligent Machine Learning Algorithm with PINIP Radial Junction Imager" Symmetry 15, no. 12: 2127. https://doi.org/10.3390/sym15122127

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop