Next Article in Journal
Recent Developments and Challenges on Beam Steering Characteristics of Reconfigurable Transmitarray Antennas
Next Article in Special Issue
Res-CDD-Net: A Network with Multi-Scale Attention and Optimized Decoding Path for Skin Lesion Segmentation
Previous Article in Journal
Sustainable Management and Policies: The Roles of Stakeholders in the Practice of Inclusive Education in Digital Transformation
Previous Article in Special Issue
Wound Detection by Simple Feedforward Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications

1
Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, China
2
MGH/BWH Center for Clinical Data Science, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA
3
Center for Advanced Medical Computing and Analysis, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA
4
Gordon Center for Medical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114, USA
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(4), 586; https://doi.org/10.3390/electronics11040586
Submission received: 12 January 2022 / Revised: 7 February 2022 / Accepted: 8 February 2022 / Published: 15 February 2022
(This article belongs to the Special Issue Machine Learning for Medical Imaging Processing)

Abstract

:
Compressed sensing (CS) and its medical applications are active areas of research. In this paper, we review recent works using deep learning method to solve CS problem for images or medical imaging reconstruction including computed tomography (CT), magnetic resonance imaging (MRI) and positron-emission tomography (PET). We propose a novel framework to unify traditional iterative algorithms and deep learning approaches. In short, we define two projection operators toward image prior and data consistency, respectively, and any reconstruction algorithm can be decomposed to the two parts. Though deep learning methods can be divided into several categories, they all satisfies the framework. We built the relationship between different reconstruction methods of deep learning, and connect them to traditional methods through the proposed framework. It also indicates that the key to solve CS problem and its medical applications is how to depict the image prior. Based on the framework, we analyze the current deep learning methods and point out some important directions of research in the future.

1. Introduction

Compressed sensing (CS) is an important problem in signal process. It can be described as reconstructing signal x from its measurement y where x R n , y R m , m < n and y is obtained in the following form:
y = A x + ε .
A R m × n defines the measuring system and ε is the noise. Reconstructing high quality images or signals has been an active area of research and holds high value in many applications, especially in medical imaging reconstruction such as computed tomography (CT), magnetic resonance imaging (MRI) and positron-emission tomography (PET). In the past two decades, traditional CS theory has been established to reconstruct x from y . Due to m < n , solving the inverse problem is not easy. Based on sparsity of x , many optimization algorithms were proposed to solve it. Though the traditional CS theory is pretty and elegant, there are still some drawbacks. For example, classic algorithms usually take a long time to solve the CS problem. Recently, deep learning—a data driven method—has demonstrated tremendous success in many fields and there is a trend to use it to solve the CS problem. Deep learning is a class of machine learning approaches that utilize cascaded layers of linear and nonlinear functions to learn the complex mapping from data. When networks go deeper with more parameters, its capability of learning features is improved, which allows the deep network to learn complex functions directly from data without human-crafted features. The core of deep learning, deep neural network, dates back to 1950s. Modern techniques, including improvements on optimization algorithm (stachastic gradient descent (SGD), rectified linear units (ReLU), batch normalization, dropout, shortcut connection et al.), more effective network architectures (convolutional neural networks (CNN), recurrent neural networks (RNN), generative adversarial networks (GAN)), the availability of large datasets and stronger computational power of hardware (GPU and parallel computing), contribute to the tramendous success of deep learning. In this review, we focus on the application of deep learning in the general CS problem and three types of medical imaging—CT, MRI and PET.
Different from some other reviews [1] which divide deep learning methods into several categories, we attempt to construct a unified framework to cover all these categories. The analysis begins with the variational model and a simple algorithm. Usually, the object of a variational model is to minimize the following function:
min x f y , A x + i = 1 K λ i R i ( x ) .
f y , A x represents the data consistency and R i ( i = 1 , , K ) are regularization terms. For simplicity, suppose that there is only one regularization term and f y , A x = y A x 2 2 , then Equation (2) can be written as follows:
min x y A x 2 2 + λ R ( x ) .
The common choice for R ( x ) is Total Variation [2] or W x 1 where W is some linear transform such as the wavelet transform. We use a simple iterative algorithm to solve Equation (3). The iterative process can be written as follows:
x k + 1 2 = arg min x y A x 2 2 + x x ( k ) 2 2 , x ( k + 1 ) = arg min x λ R ( x ) + x x k + 1 2 2 2 .
This algorithm contains two steps. By geometric analysis, the first step moves x to a position closer to the hyperplane y = A x and the second step moves x to a position with lower value of R ( x ) . If we regard the regularization term as a depiction of the manifold of signals, the iterative algorithm derives a solution by alternatively moving x to the hyperplane and the manifold. The movement of x is shown in Figure 1.
From the Bayesian view, we can understand the role of regularization terms more clearly and figure out what the solving algorithm do. Suppose ε N 0 , σ 2 I and the prior distribution of x is p x . Then we derive the logarithmic posterior probability of x as follows:
log p x y = y A x 2 2 + λ log p ( x ) .
Here, for simplicity, the coefficient of y A x 2 2 is neglected. If we apply a simple first-order gradient method to maximize the posterior probability, the iteration will be in the following form:
x ( k + 1 ) = x ( k ) + η A H y A x ( k ) + η λ log p x ( k )
where A H is the conjugate transposition matrix of A and η is the step length. It is easy to verify that η A H y A x ( k ) represents a direction toward the data consistency hyperplane y = A x and η λ log p x ( k ) toward higher prior probability. The geometric interpretation is illustrated in Figure 2. We can see the similarity between the variational model and the Bayesian model. In other words, regularization terms correspond to the representation of logarithmic prior distribution of x . Thus, we have the following conjecture: the solving algorithm of the CS problem is to search a solution that is in the intersection of the data consistency and the prior information. It contains two parts. One is a “projection” operator to the image prior and the other one is to the data consistency.
Based on the geometric analysis of optimization algorithms, we can define a unified framework for solving the CS problem. Since the typical signals in CS are images and three applications reviewed here are medical imaging, we only discuss image signals through the review. Let M image be the manifold representing the image prior and M dc be the solution space of the data consistency. We define P image as a transform that projects x toward M image and P dc as one projecting x toward M dc . We claim that a solving algorithm satisfies the framework F if it is composed of P image s and P dc s. Sometimes, P image can be further decomposed into three transforms:
P image = V x P x U x ,
where U x : R n S , P x : S S , V x : S R n . U x transforms an image to a defined space S; P x defines the "projection" operator in S and V x transforms the result back to the image space. This decomposition means that the image prior can be depicted in space S rather than R n . For example, suppose the regularization term in the variational model is W x 1 where W is a wavelet decomposition operator. Then U x is the wavelet transform, V x is the inverse transform and P x can be the soft-thresholding function. We find that the solving algorithms for the variational model or the Bayesian model satisfy F . Though F is very simple, it is surprising that almost all deep learning methods solving the CS problem or its applications also satisfy this framework. Thus, the framework F provides a perspective to analysing solving algorithms. In addition, since CS belongs to inverse problems, this framework can be expanded to other inverse problems such deblurring, inpainting, etc. as long as we choose a feasible P dc .
Our main contributions are as follows:
  • We proposed a framework which unifies traditional iterative algorithms and deep learning approaches for CS reconstruction and its medical applications.
  • We reviewed many works on reconstruction of CS, CT, MRI and PET, and analyzed them based on the proposed framework.
  • Through the proposed framework, we built relationship between different reconstruction methods of deep learning and indicated that the key to solve CS problem and its medical applications is how to depict the image prior.
In later sections, we also divide deep learning methods into different categories. Nevertheless, the emphasis is to illustrate how these categories match the framework F . This review is organized as follows. Section 2 describes deep learning methods used in general CS. Some works for CT reconstruction are reviewed in Section 3. Section 4 surveys recent deep learning methods for MRI reconstruction. Then, we provide some deep learning approaches for PET reconstruction in Section 5. Finally, we compare these methods and discuss future directions in Section 6 and concludes the review in Section 7.

2. Deep Learning Methods for Compressed Sensing

2.1. Overview

In the general CS reconstruction problem, usually x is a natural image and A is a Gaussian random matrix. In this section, we divide deep learning approaches into five categories and analyze how each one matches framework F .

2.2. Model-Based Methods with Learnable Parts

The first category is model-based methods with learnable parameters. These methods may be traced back to learned iterative shrinkage and thresholding algorithm (LISTA) [3]. Convolutional Neural Networks (CNN) or other neural networks are not used. Instead, some pre-fixed parameters or functions in traditional algorithms are learned from training data. Generally speaking, through loss function and back-propagation method, these algorithms can be regarded as a trainable network [4,5,6,7]. Suppose the traditional algorithm is Alg ( · ; θ ) where θ is the pre-fix parameters. Then the reconstruction of it can be written as Alg ( y ; θ ) where y is the measurement. Let ( x i , y i ) i = 1 N is the training set and L is the loss function, then the training process is to optimize over θ by min θ i = 1 N 1 N L ( Alg ( y i ; θ ) , x i ) where x i is the training label. Therefore, Alg ( y i ; θ ) can be regarded as a trainable network with learnable parameters θ . The purpose of applying data-driven scheme is diverse. Some are to reduce computation cost, some to ascertain best parameters and some to make regularization terms closer to the image prior. Since the overall form is not changed and the original algorithm itself satisfies the framework F , these methods still consist of P image s and P dc s and therefore match the framework F .
The authors of [4] proposed to replace the soft thresholding function in iterative shrinkage and thresholding algorithm (ISTA) by other learnable non-linear functions. They used cubic spline functions as basis functions and learned the weights c k . The alternative function has the following form:
φ ( z ) k = K K c k ψ z Δ k ,
where ψ is the cubic spline function, Δ is the granularity parameter and K is the number of basis functions. Given fixed T iterations, L2 norm loss between final reconstruction results and real images was used to train the weights c k . In addition, the authors of [5] not only used Equation (8) but also trained the step length. Similarly, Gaussian kernel functions were used as basis functions to replace the proximal operator in ISTA [8]. The shrinkage function ψ t ( u ) is written as follows:
ψ t ( u ) = k = 1 K c k t ϕ k ( u ) , where ϕ k ( u ) = u e ( k 1 ) u 2 2 τ 2 .
ϕ k is the Gaussian kernel function and K is the number of basis functions and t is used to represent different steps. To reduce number of learnable parameters, the authors of [6] proposed to employ linear expansion of thresholds (LET) to substitute the soft thresholding function. Besides, they considered fast ISTA (FISTA) instead of ISTA.
The authors of [7] proposed a novel network, ISTA-Net, which replace the linear transform in the regularization term by a two-layer neural network. In original algorithm, the second step of iteration is a proximal operator which has the following form:
prox λ ϕ ( r ) = W soft ( Wr , λ ) .
They used F ( · ) and F ˜ ( · ) (two-layer neural networks) to substitute W and W . Since W W = I , the constraint of F ˜ F = I is added to the loss function. It has the following form:
L total ( Θ ) = L discrepancy + γ L constraint = 1 N b i = 1 N b x i N p x i 2 2 + 1 N b i = 1 N b k = 1 N p F ( k ) F ( k ) x i x i 2 2 ,
where N b is the amount of data and N p is the number of iterations. The input of the network is an initial reconstructed image. Based on ISTA-Net, they considered residual learning and proposed a modified version ISTA-Net+.
A recent work [9] proposed to substitute the convolutional operator in transform learning algorithm by a learnable convolutional layer with 3 × 3 kernels. The object function is as follows:
min x , α k y Φ x 2 2 + η k K W k x α k F 2 + J α k .
The iterative process is shown as follows:
x = argmin x y Φ x 2 2 + η k K W k x α k F 2 , α k = argmin α W k x α k F 2 + J α k .
The first sub-problem can be solved by gradient methods:
x ( t + 1 ) = x ( t ) δ Φ Φ x ( t ) y + η k K W k W k x ( t ) α k ( t + 1 ) .
Under some assumptions, Equation (14) is simplified as a residual form:
x ( t + 1 ) ρ x ( t ) + δ x ( 0 ) + γ x ( t + 1 / 2 ) ,
where x ( t + 1 / 2 ) = k K W k α k ( t + 1 ) . It is the output of the convolutional layer. Then, the unrolled iterative algorithm is changed to a network. Moreover, the measuring matrix is replaced by a convolutional layer with m channels, L × L kernel size and s stride. The initial reconstruction is computed by another convolutional layer. All the convolution parameters are learnable.
There are some other works that belong to this category. The authors of [10] proposed Iterative Firm Thresholding Algorithm (IFTA) to solve general inverse problem and set most parameters to be learnable. In [11], the weights of proximal operator is obtained by training. For low-rank tensor factor analysis approach, the authors of [12] used neural networks to substitute the matrix computation.

2.3. Neural Networks as Image Projections

The second category is to directly use neural networks (or some deep learning modules) as the P image . In this category, an initial reconstructed image is needed. In some works, the initial reconstruction is contained in the beginning part of neural networks. At first, the reconstruction model only contained one P image and no P dc , just like a denoising model. Later, more sophisticated networks were proposed and P dc was included as one layer of the model. When more than one P image and P dc occur in the network, it has an unrolling form similar to traditional iterative algorithms such as alternating direction method of multipliers (ADMM), ISTA and denoising approximate message passing (D-AMP). It is worth noting that when P image is represented by a neural network, the image prior is hidden. In this category, networks can be regarded as substitutes for the original proximal operator in iterative algorithms. Different algorithms lead to different form of networks. Generally speaking, most improvements are about network architecture and loss function design.
The authors of [13] proposed to use a three-layer fully-connected network to reconstruct image from measurements. The input of network is measurements and the output is reconstructed images. Since fully-connected layers are used, the network is trained with 32 × 32 patches to reduce parameters. Correspondingly, the measurements are obtained from image patches. Later, a convolutional neural network was applied in the same manner [14]. The first layer of the CNN is still a fully-connected layer to transform the measurement to image space. In this work, patches with size of 33 × 33 were used for training. In addition, block-matching and 3D filtering (BM3D) [15] is exploited as post-process to overcome the blocky artifacts when reconstructed image patches are pieced together to form the whole image. In [16], the first fully-connected layer of the CNN is replaced by the transposition of measuring matrix. The CNN proposed by [17] contains not only the reconstruction part but also the measuring part. One convolutional layer with n B channels, B × B kernel size and B stride plays the role of measuring. It is followed by a convolutional layer with B 2 channels and 1 × 1 kernel size which is used to reconstruct image initially. The output is then reshaped to the original image size. In fact, such measuring and initial reconstructing manner is equivalent to block CS. However, it makes it possible that the whole image can be fed into the network.
Residual structure [18] was applied to reconstruct image [19]. The first layer of the network is a fully-connected layer to transform measurements into image space. The following part contains several residual learning blocks. Similar to [14], patches are used for training and BM3D is also exploited to remove blocky artefacts. The training scheme is composed of two steps. Firstly, the fully-connected layer is trained using mean squared error (MSE) loss. Then the whole network is trained in an end-to-end manner. The authors of [20] proposed a similar method, but the measuring matrix is replaced by one fully-connected layer and also trained together with other part of network. In [21], the BM3D module is substituted by another residual convolutional block which is the combination of 11 × 11 , 1 × 1 , 7 × 7 convolutions and ReLU functions. Different from [19], in [22] the measuring and initial reconstructing parts are convolutional and deconvolutional layers, respectively, so as to reconstruct the whole image instead of patches.
The choice for loss functions is also explored in some works. Besides popular MSE (L2 norm) loss, adversarial loss [23] is used when training networks [24]. It was proposed firstly to train generative adversarial networks (GAN). The basic network architecture is an analogy to the reconstruction part in [21]. Perceptual loss is exploited in [25] and structural similarity (SSIM) loss is applied in [26]. All these loss functions are used to enhance the quality of reconstructed images.
More works focus on how to design the network architecture to achieve better reconstruction performance. A two-branch network was proposed by [27]. One branch utilizes dense connection structures and the other one consists of residual blocks. Random sampling scheme and fully-connected sampling scheme are all considered. Since it is based on block CS, BM3D is also used to remove blocky artefacts after patch reconstruction. The authors of [28] proposed a pyramid-structured adversarial network. In general CS problem, reconstructed images have a fixed resolution. As long as the number of measurements is insufficient, the reconstruction quality is unsatisfied. The idea of the pyramid network is that the resolution of reconstructed images depends on the number of measurements. A low-resolution image is reconstructed from fewer measurements while high resolution ones are reconstructed from more measurements. Different levels of resolution correspond to different sub-networks. The input of sub-networks is the reconstructed image from last level and measurements.
Scalable sampling rates are considered in [29] and SCSNet was proposed. Measurements are divided into groups and used as reconstructed information for different scales. One group is used to reconstruct the low frequency part of images which corresponds to the basic layer in network. Others are used to reconstruct the high frequency part corresponding to enhanced layers (EL). Measurements are obtained from image patches by a non-overlapping block convolutional layer. After initial reconstruction, a deep reconstruction network is applied to reconstruct the whole image. In this work, the MSE loss function is applied to both initial reconstruction and final reconstruction.
A recent work exploited the idea that reconstructed signals can be decomposed into two orthogonal parts [30]. One is in the null space of the measuring matrix H and the other is in the pseudo-inverse space. Suppose the measurements satisfy y ε = H x + ε . x is decomposed by x = P r ( x ) + P n ( x ) where P r H H and P n I D H H . Then we can derive that x = H y ε + H ε + P n ( x ) . The network consists of two parts which is used to reconstruct the two signal components, respectively. The authors of [30] considered two forms of architectures.
Multi-scale structures were utilized in [31]. There are three branches of sub-networks with different convolutional kernel sizes to extract information of different scales. All the sub-networks have residual blocks and non-local layers which are helpful for global information extraction. At the beginning of training, three sub-networks are trained, respectively, and non-local layers are neglected. Finally, the whole network is trained in an end-to-end manner.
The works reviewed above are all about how to design a neural network as the P image , and no P dc is used. Some works generalize this approach to use neural network many times and combine it with traditional iterative algorithms to form an unrolling architecture. Since the P dc is contained in iterative algorithms, the whole network is composed of many P image s and P dc s. Usually, the P dc retains the original form. In other word, in this approach neural networks substitute the original P image s of the iterative process. The role of P image s in unrolling methods is, in essence, the same to those that only use one P image . Therefore, any network design mentioned above is also applicable. Some works train the P image network beforehand, while others train the unrolled network in an end-to-end manner. As for the concrete form of P dc , gradient computation is used in some works while the proximal operator is used in others.
The authors of [32] proposed to train a projection network to replace proximal operators in the iterative algorithm. The purpose is to solve all the inverse problem, including CS reconstruction. The projection network plays the role of P image and is trained beforehand. It contains an auto-encoder P and two discriminators, D and D . The input of auto-encoder is a clean image or a perturbed one which is obtained by adding Gaussian noise. D is used to discriminate the outputs of P while D is for the encodes of P . The loss function has the following form:
min θ P x M , v f ( x ) λ 1 x P ( x ) 2 + λ 2 x P ( v ) 2 + λ 3 v P ( v ) 2 λ 4 log σ D E ( v ) λ 5 log ( σ ( D P ( v ) ) ) ,
where x is a clean image, v is the perturbed one and E is the encoder of P . In this work, ADMM algorithm is used and the trained projection network substitutes the first step of iteration. The authors of [33] also proposed to use a neural network to represent the proximal operator in ADMM algorithm. However, they utilized a denoising CNN with residual structures and different noisy level are tested to attain the best performance. Similarly, the proximal operator in proximal gradient method is replaced by a neural network in [34]. The authors of [35] proposed to use a neural network as denoising model in D-AMP algorithm. The modified D-AMP algorithm has the following form:
b t = z t 1 div D σ ^ t 1 x t 1 + A H z t 1 m ,
z t = y A x t + b t ,
σ ^ t = z t 2 m ,
x t + 1 = D σ ^ t x t + A H z t .
where D σ ^ t 1 is the neural network. The D-AMP algorithm was also applied to block CS reconstruction in [36]. The denoising model, i.e., P image is a DnCNN [37]. For efficiently sampling, the sampling rate of patches depends on the salient value. Patches with the same value are measured by the same measuring matrix which, specifically, is a convolutional layer. P dc is computed for patches while P image is computed for the whole image.
The authors of [38] proposed to treat a neural network as a regularization term. The model is written as:
x rec = arg min x A ( x ) b 2 2 data consistency + λ N w ( x ) 2 regularization .
where N w ( x ) = I D w ( x ) = x D w ( x ) and D w ( x ) represents a neural network. Then the unrolling architecture can be derived as follows:
x n + 1 = arg min x A ( x ) b 2 2 + λ x z n 2 ,
z n = D w x n .
The first step corresponds to the P dc to keep data consistency. While the neural network D w is the P image . To reduce parameters, weights of the denoiser in all iterations are shared. The training scheme contains two stages. In the first stage, only one iteration is trained. In the second stage, all the iterations with shared weights are trained together. A similar approach was proposed in [39] and the authors hold a viewpoint that residual structure is feasible to represent prior. The authors of [40] proposed an unrolling network based on a primal-dual algorithm where proximal operators are replaced by a three-layer network with PReLU activation functions. An extra gradient method was unrolled in [41] and Nesterov’s accelerated gradient method was utilized.
The authors of [42] proposed a Network-based PGD (NPGD) method to reconstruct images from CS measurements. The P image in this work is not a denoising model, but a composition of a GAN and its inverse network. Firstly, a trained GAN is used for depicting image prior. The generator is denoted by G and its inverse network G is trained to project a image signal to the latent space of G. Thus, G G plays the role of P image . The following loss function was proposed to train G ,
L ( θ ) = E z , ν G G θ ( G ( z ) + ν ) G ( z ) 2 + E z , ν λ G θ ( G ( z ) + ν ) z 2 .
Based on unrolling networks, some works focus on improvements of denoising models. The authors of [43] divided the model into three sub-models which are based on MWCNN [44]. Each one deals with different levels of noise and their average output is used finally. In addition, the input of sub-models is expanded in channels and each channel keeps identical. Usually, the input of the P image is a corrupted image. However, in [45] image is decomposed into several combinations of high frequency parts and corresponding low frequency ones. Only high frequency ones are input. After denoising, the clean high frequency parts are added to corresponding low frequency one. Finally, the average of different combination is the output of the P image . Frequency decomposition is realized through minimizing an object function consist of a total variation with different coefficients. The coefficients control the frequency decomposition.

2.4. Latent Variable Search of Generative Models

The third category is the latent variable search of the generative model. The basic idea is simple. Firstly, a generative model, such as GAN, is trained. Its output represents the image prior manifold. Then, minimize a loss function, which usually corresponds to the data consistency, by searching the latent variable. Generally speaking, the object of CS reconstruction is to find a best image x . However, in this category of methods the search of x is replaced by the search of latent space variable. Suppose the trained generative model is G, latent variable is z , and data consistency is represented by y A x 2 2 . Then the optimization problem is as follows:
min x y A x 2 2 , s . t x = G ( z ) .
It can also be rewritten as follows:
min z y A G ( z ) 2 2 .
When the solution z * is obtained, reconstruction result is derived by G ( z * ) . At first look, it is hard to verify that this method also satisfies framework F . Suppose that we use a simple first order gradient method to solve Equation (26), we have the following decomposition for each iteration by the chain rule:
z ( k + 1 ) = z ( k ) η A G z ( k ) y 2 2 z ( k )
= z ( k ) η D ( k + 1 ) r ( k + 1 ) ,
where r ( k + 1 ) = A G z ( k ) y 2 2 G z ( k ) , D ( k + 1 ) = G z ( k ) z ( k ) . In fact, the generative model represents M image , and the composition of P image and P dc can be represented by P image P dc = V x P x U x where U x = G 1 ( x ) , P x = z η D r and V x = G ( z ) . P d c is implied by the loss function and is hidden in the derivative computation of r . Actually, r corresponds to P d c . Figure 3 shows the movement of x . After computing r , the direction is limited to the latent space by D r . Then through the generative model, the limited direction corresponds to the movement of x and P image is realized. Thus, this category also satisfies framework F .
This approach was firstly proposed in [46] and two theorems about the error upper bound are given. An improvement was proposed by the authors of [47]. A sparse item is added to correct the reconstruction. The object function has the following form:
min z , v v 0 ,
s . t . A ( G ( z ) + v ) = y .
Then CS problem is solved by minimizing a non-constraint object function as follows using a first-order gradient method:
min z , v v 1 + λ A ( G ( z ) + v ) y 2 2 .
Here, zero norm is replaced by L1 norm. Another improvement in [48] is that the latent variable is also optimized when training the GAN model. In other word, a set of latent variable { z ^ ( 1 ) , z ^ ( 2 ) , , z ^ ( s ) } is trained to satisfy that y ( i ) = A G ( z ^ ( i ) ) . When training data is insufficient, another discriminator is applied to discriminate measurements besides the usual image discriminator. Auto-encoders and generative models are combined in [49]. Auto-encoders tend to effectively extract low-frequency structure of image while losing details. GANs are good at generating images with fine details but may cause global corruption. Thus, the fitting in measurement in Equation (26) is substituted by encode fitting. In addition, z 2 2 is added as a regularization term.
Instead of first-order gradient method, ADMM algorithm is also used to solve Equation (26). In [50], suppose there is a regularization of z denoted by H ( z ) . Then the object function is min x , z y Φ x 2 2 + λ H ( z ) , s . t . x = G ( z ) . The iteration has the following form:
x ( k + 1 ) = Φ T Φ + ρ I 1 Φ T y + ρ G z ( k ) μ ( k ) ,
z ( k + 1 ) = arg min z λ H ( z ) + ρ 2 x ( k + 1 ) G ( z ) + μ ( k ) 2 2 ,
μ ( k + 1 ) = μ ( k ) + x ( k + 1 ) G z ( k + 1 ) .
To solve the second step, a fully-connected network G proj was proposed in [50]. It has to be trained using pairs of ( x ˜ , z ) where x ˜ is a noisy signal represented by x ˜ = G ( z ) + ε . Another trick in training the GAN is that each latent variable z is split into code-words c and “random-noise-like” variable γ , which is inspired from InfoGAN [51]. c is used to control the semantic information and γ controls variation. A loss function that maximize the mutual information between c and G ( z ) is included. The authors of [52] proposed a new training strategy combining meta learning and generative model to accelerate the search of latent variable.

2.5. Neural Networks Based Probability Models

The fourth category is to use neural network to represent prior distribution of images and maximize the posterior probability. It is one of Bayesian models and the projection direction of the P image and the P dc are related to x log p ( x ) and x log p ( y | x ) . Thus, this category satisfies Framework F .
RIDE model was proposed in [53]. It combines a LSTM [54] model and Mixture of Conditional Gaussian Scale Mixtures as image prior distribution. Then the gradient method is used to solve the posterior distribution. Later, in [55] a PixelCNN [56] was applied to represent image prior. Its model has the following form:
p ( x ) = p x 1 , x 2 , , x n 2 = i = 1 n 2 p x i | x < i ) .

2.6. Unsupervised Methods

Last category is unsupervised method. When there is no real image dataset, it is hard to depict image prior. In [57], deep image prior (DIP) method was proposed. It was used to solve some inverse problem not including CS. Later, DIP was applied to reconstruct image from compressed measurements. Most of current unsupervised methods are based on it. The basic idea is to use an untrained generative model and minimize the loss function of data consistency over network parameters with fixed input z . DIP method is similar to the category discussed in Section 2.4. However, M image is represented by an untrained network itself instead of a trained generative model. In other words, image prior is depicted by the output of generative model with fixed latent variable and learnable network parameters. Searching in parameter space is analogue to searching in latent space. Thus, similar analysis of Section 2.4 can be used to verify that this category also satisfies framework F .
The authors of [58] proposed to applied DIP method to solve CS reconstruction. A regularization term is added to loss function which has the following form:
arg min w y A G ( z ; w ) 2 + R G ( z ; w ) , w ; λ T , λ L ,
where R G ( z ; w ) , w ; λ T , λ L = λ T T V ( G ( z ; w ) ) + λ L ( w μ ) T Σ 1 ( w μ ) . μ and Σ are the mean and covariance matrix of network parameters estimated by a few data. Total variation regularization was used in [59] to help reconstruction. In [60], semi-supervised learning was discussed. In Section 2.4, training a generative usually need a great deal of data. The authors of [60] proposed a strategy to make a trade-off. In pre-train stage, network parameters and latent variables are trained simultaneously with a combination of image L2 loss and kernel loss. The latter has the following form:
min θ , z 1 , , z S 1 S 2 i i k G z i ; θ , G z i ; θ + 1 S 2 j j k x j , x j 2 S 2 k G z i ; θ , x j ,
where z i is the latent variable, θ is network parameters and S is the number of training sample. In the reconstruction stage, latent variable is first optimized and then together with network parameters.

2.7. Discussion

In this section, we further explained different categories of deep learning methods by our framework, especially latent variable search of generative models. We can observe a trend from simple networks to complex and bigger ones. Among these methods, cascaded networks, which serve as image projections perform best. While generative model or probability model based methods are less comparable due to the unsatisfied performance of generative models and probability model. However, they still have potential to be improved in the future as more powerful generative models and probability models are proposed.

3. Deep Learning Methods for Computed Tomography

3.1. Overview

We mainly discuss sparse-view or limited angles CT reconstruction in this section. All the works reviews here belong to the five categories of Section 2. Therefore, they must satisfy framework F . The emphasis is to illustrate how they design P image and P dc . It is worthy of mention that the initial reconstruction is obtained by the FBP algorithm.

3.2. Model-Based Methods with Learnable Parts

There are few works belonging to this category. The authors of [61] proposed to applied variational network to low-dose CT reconstruction. Fields of experts are used as regularization term of variational model shown as follows:
R c ( u ) = 1 , ϕ c K c u ; W c
where u is a CT image, ϕ c is a linear interpolation and K c is a convolution. K c and W c are learned by training. The network architecture corresponds to unrolling the first-order gradient method:
u t = u t 1 K c ϕ c K c u t 1 ; W c λ c A ( A u t 1 d ) .
JSR-net [62] was proposed to solve limited angle CT reconstruction. It unrolled the ADMM algorithm for JSR-model. The computation of two inverse matrices and the thresholding function in JSR-model are replaced by neural networks. The former is substituted by a three-level DenseNet with LM-ResNet structure and the latter by a three-layer convolutional network. The object function of JSR model has the following form,
min u , f F ( u , f , Y ) + λ 1 W 1 u 1 , 2 + λ 2 W 2 f 1 , 2 ,
where
F ( u , f , Y ) = 1 2 R Γ c ( f Y ) 2 + α 2 R Γ ( P u f ) 2 + γ 2 R Γ c ( P u Y ) 2 ,
and Γ c represents sampling angles.

3.3. Neural Networks as Image Projections

Most works of CT reconstruction using deep learning methods belong to this category. The authors of [63] applied a fully-connected network to refine the middle result of traditional iterative algorithms. A three-layer CNN was used to low-dose CT reconstruction task in [64,65]. The input of network is initial reconstruction of the FBP algorithm. The authors of [66] proposed to applied a U-Net for reconstruction. Residual structures were considered in [67]. Later, the authors of [68] proposed to add bypass connections and utilize the Haar wavelet as down-sampling and up-sampling to improve the reconstruction.
Almost all of methods employ L2 or L1 norm loss of image. Some works also apply other type of loss functions. Perceptual loss was used in [69]. Adversarial loss function was exploited in [70]. A discriminator was used to help to refine the details of reconstruction by adversarial training.
Some works change the object of P image . It means that there are explicit U x and V x . In [71], input of the U-Net is the result of wavelet decomposition of the initial reconstruction image which purpose is to utilize multi-scale information. In other word, U x and V x is related to wavelet decomposition and synthesis. Using sinogram measurements as network inputs was proposed by [72]. For different angles of view, corresponding sinogram was expand to a image by back projection and these images were stacked to form a tensor. Then it was used as input of a 15-layer CNN. In this work, U x is the process of sinogram and V x is merge into P x . In [73], interpolated sinogram was used as the input of a U-Net and the output is the accurate sinogram. When output of network is obtained, FBP algorithm is applied to compute the final reconstruction. In this method, U x is the transform from initial image to sinogram space, V x is executed by the FBP algorithm and P x is the U-Net. Thus, the projection operator is executed in sinogram space. This is an example illustrating a difference between medical application and general CS problem. In fact, sinogram is the measurement y and there exists a transform and its inverse between measurement space and image space. We can also regard the method in [73] as a projection model in measurement space. Similar to [73,74] used a U-Net to reconstruct under-sampled sinogram. Besides, a discriminator and adversarial training was exploited in this work. The input of the discriminator is the sinogram with limited angles and full-size output of generative.
All the works mentioned above only consist of one P image and no P dc . Similar methods can be seen in [75,76,77,78,79,80,81]. Next, deep learning methods with more than one P image and P dc will be reviewed.
The authors of [82] considered a regularization term of Fields of Experts and used a simple first-order gradient method to solve the object function. It has the following form:
x t + 1 = x t λ t A T A x t y + k = 1 K G k t T γ k t G k t x t ,
where k = 1 K G k t T γ k t G k t x t is related to Fields of Experts. This term was replaced by a three-layer CNN which plays the role of the P image . Since Equation (42) is in an iterative form, the network contains many P image s and P dc s.
The authors of [83] proposed to unroll the ADMM algorithm and added a regularization term about sinogram to original object function. Thus, there are two types of P image s. The object function is shown as follows:
min x , y 1 2 y y ^ Σ y 1 2 + 1 2 A x y Σ x 1 2 + λ R y ( y ) + γ R x ( x ) .
Though there is an explicit transform relationship between sinogram y and image x , in the optimization task they are split to exploit λ R y ( y ) . The sinogram regularization is R y = 1 2 j m N j ω j m y j y m 2 . When the iteration is unrolled into a network, a ResNet was applied to deal with γ R x ( x ) . Besides, L2 norm loss with weights (indicated by Σ y and Σ x ) was used for y and x . In [84], unrolling the ADMM network was also used. However, proximal operator of the regularization term was substituted by a U-Net. In [85], a more complex object function is solved by unrolling the ADMM network. The ADMM iteration contains four proximal operators and they were all replaced by three-layer CNNs. That is to say, both P image s and P dc s are represented by neural networks.
The authors of [86] used a denoising auto-encoder with soft-thresholding function as P image and solved the P dc by FISTA. In each stage, a cleaner image z is obtained by the denoising model and FISTA is used to keep data fidelity of z . Because FISTA is not unrolled into a network, the parameters in the P image cannot be trained in an end-to-end manner. A stage-wise training scheme was proposed to solve the problem.
In [87], a proximal forward backward splitting algorithm was unrolled into a network to reconstruct CT image. It is similar to ISTA network and the proximal operator is replaced by a CNN. However, instead of last iteration result, all iteration results before are used as the input of CNN in next iteration. In addition, the pseudo-inverse of measuring matrix rather than transposition is used to compute the P dc .
Scale invariant property was exploited in [88]. It is combined with the unrolling network. Specifically, the granularity in each iteration becomes finer and in last iteration the original full measuring matrix is used. The iteration has the following form:
f i = Λ θ i f ˜ i , D i f ˜ i ; g , f ˜ i + 1 = τ i + 1 f i ,
where D i f i ; g : = A i * A i f i π i ( g ) , Λ θ i corresponds to the P image and τ i + 1 is upsampling operator. The multi-scale idea is similar to multi-level structure in U-Net. Thus, the unrolling form is represented by a U-Net.
In [89], both CNNs and traditional algorithms are used to reconstruct CT image. The methods are used alternatively to improve reconstruction. The recurrent scheme means that there are two types of P image (CNNs and the regularization term in iterative algorithms) and one type of P dc . FBPConvNet is chosen as the neural network structure and PWLS-EP or PWLS-ULTRA is the choice for the iterative algorithm. Similar to [86], the training scheme is a stage-wise process.
Other deep learning method in unrolling form can be seen in [90,91] and etc.

3.4. Discussion

Most works on CT reconstruction are very close to solve a denoising problem. We found that there are few works that focus on latent variable search of generative models or probability model due to the complexity of the measurement matrix. How to design effective algorithm to combine generative models or probability models and CT reconstruction is an interesting direction in this area.

4. Deep Learning Methods for Magnetic Resonance Imaging

4.1. Overview

In this section, we focus on under-sampled MRI reconstruction which is an important application of CS reconstruction. Some properties of MRI reconstruction distinguish it from other CS problem. Firstly, the image is in the field of complex number. In MRI reconstruction, the measurement is called k-space coefficient which is, in fact, the result of the Fourier transform of image. Thus, the measurement and image are represented by complex number. The magnitude of image is used to show the image. For traditional iterative methods, the operations of real number are easy generalized to complex number. However, how to deal with complex number for neural network is a problem since it is based on tensor operations. For most works using deep learning methods, complex numbers are represented by two-channels tensors, i.e., any x C n is regarded as in R 2 n . This treatment is equivalent to regard complex number images as two-channel real number images and all the computation is based on real numbers. Another kind of method is to simultaneously keep the complex number operation and use two-channel representation [92]. Secondly, there is a special imaging method called parallel imaging which makes the linear model more complex. In parallel imaging, several coils are utilized and each one corresponds to a k-space measurement, respectively. If every coil is under-sampled, reconstruction images will be still of high quality while reducing scan time. However, the acceleration of this method is limited. Combining with CS reconstruction can further accelerate the scan. In addition, for each coil there is a sensitive matrix differing in every scan and relating to the k-space measurements. The forward model has the following form:
y i = A S i x = M F S i x , i = 1 , 2 , , c ,
where A represents the Fourier transform F with under-sampled mask M , S i is sensitive matrix for the ith coil and c is the number of coils. Since sensitive matrices are not fixed parameters, they have to be estimated when reconstruction. A common approach is to estimate sensitive matrices beforehand using SENSE [93] or other algorithms and then regard A S i as a fixed measuring matrix. Therefore, each coil has its own data consistency. This is the main difference between single-coil imaging and parallel imaging.
Besides, it is worthy of mention that the P dc in MRI has a very popular form as follows:
y ^ j = F ( N ( x ) ) j , if j Ω , F ( N ( x ) ) j + λ y j 1 + λ , if j Ω .
where Ω is the sampled position, N ( x ) is the current reconstructed result and F is the Fourier transform. λ is the weight to control the extent of data consistency. When λ = , original sampled k-space coefficients will be retained.
In later sections, deep learning methods will be also divided into several categories and the criterion is the same to Section 2. Each categories will be further divided into non-parallel imaging and parallel imaging sub-categories if necessary.

4.2. Model-Based Methods with Learnable Parts

4.2.1. Non-Parallel Imaging

In [94], the original objection function is as follows:
x ^ = arg min x 1 2 A x y 2 2 + l = 1 L λ l g D l x .
Based on ADMM algorithm, the linear transform D l in regularization terms was replaced by learnable convolutions and the shrinkage function in iterations was substituted by a learnable piece-wise linear function. step lengths is also learnable parameters. Later, the authors proposed another form of the ADMM network in [95]. Similar to [94,96] also unrolled a ADMM network. Because the noise model is supposed to be symmetric α -stable, therefore L1 norm loss is adopted. In practical terms, a smoothing term is used to replace L1 norm. IFR-CS model was proposed in [97] which network architecture is based on [7]. Besides data consistency and proximal operator of the regularization term, a refine step was added in the iterations.

4.2.2. Parallel Imaging

The authors of [98] proposed a variational network which is based on the Fields of Experts model. The regularization term in this model R ( u ) is i = 1 N k Φ i K i u , 1 . The first-order gradient method is used to solve the original object function which has the following form:
u t + 1 = u t i = 1 N k K i t Φ i t K i t u t λ t A * A u t y , 0 t T 1 .
All the parameters are learnable including K , Φ i and λ . Φ i t is represented by Gaussian radius basis functions. As for sensitive matrices, they are estimated by ESPIRiT [99] algorithm beforehand. The authors of [100] considered parallel imaging and their method is similar to [94]. There are also some works such as [101,102,103] that can be put under this category.

4.3. Neural Networks as Image Projections

4.3.1. Non-Parallel Imaging

The authors of [104] may be the first to apply deep learning in MRI reconstruction. Their method is to train a three-layer CNN beforehand and use the network to reconstruct image. Three ways were proposed in [104]. The first one is to minimize the following object function:
C A H y ; Θ ^ x 2 2 + λ y A x 2 2
where C is the trained network. C plays the role of the P image and the minimization of Equation (49) corresponds to P dc . The second one is to add an extra regularization to Equation (49). In the third way, the output of C A H y ; Θ ^ is used as the initial value for a traditional CS reconstruction algorithm.
The authors of [105] proposed to reconstruct the magnitude and phase of image, respectively. The network architecture is a U-Net with global residual learning. Since the value of phase in noisy district is random and meaningless, magnitude network is trained first to ascertain ROI and phase network is trained only in ROI.
The authors of [106] proposed to use conditional GAN to reconstruct MRI whose generative network is U-Net. This method is, in essence, adding adversarial loss function to train the P image . Besides MSE loss of images and adversarial loss, perceptual loss is also used. The authors of [107] applied SSIM loss to dynamic MRI reconstruction. Later, in [108], MSE loss of k-space is added. The authors of [109] proposed to utilize dense connection structure in the bottleneck part of U-Net. The authors of [110] proposed to use adversarial loss function in LSGAN [111]. In addition, the weighted average of L1 norm and L2 norm loss function is utilized. To make training stable, the weight of adversarial loss is set to zero at the beginning of training. Some works modified the structure of U-Net to improve the reconstruction. In [112], convolutions of different sizes were exploited to extract multi-scale information. The features extracted from different convolutions are fused to be the input of the next layer. In [113], dilated convolutions were utilized and residual learning structure was added in the bottleneck of the U-Net. In [114] two U-Nets with residual structures were connected sequentially as the generative network.
Since P dc is easy to implement, many works whose method contains only one P image also add one P dc to correct the reconstruction and keep data consistency. For example, ref. [115] proposed to correct k-space coefficients after using a U-Net to reconstruct images. Some works attempt to reconstruct both image and k-space coefficients. The authors of [116] proposed to use a residual U-Net to reconstruct k-space coefficients and another U-Net for images. These two networks are connected by the inverse of Fourier transform. Reconstructing k-space coefficients is similar to reconstruct sinogram of CT, which is discussed in Section 3. Thus, it corresponds to a P image which is defined in k-space. The authors of [117] proposed to employ four networks for reconstruction which is named by KIKI-net. Two is for images and two for k-space coefficients. The order is k-space, image, k-space and image (KIKI). For each image network, P dc is added to guarantee data consistency and connect adjacent networks.
There are also many works utilize more than one P image and P dc which form an unrolling network. Some works called it cascade structure because it is not necessary to be derived from an iterative algorithm. However, they are similar in essence because of the alternate order between P image s and P dc s. The authors of [118] proposed a network architecture where CNNs and data consistency are connected alternately. Since the reconstruction object is dynamic MRI, multi-frame images are trained simultaneously and 3D convolution is used. In [119], the output of each CNN in the cascade networks are concatenated at last and a convolutional layer is used to obtain the final reconstruction. The authors of [120] combined neural networks and traditional ADMM algorithm. The training and inference process are both in an ADMM iteration form. The authors of [121] proposed to unroll the Chambolle–Pock algorithm and P dc is also replaced by a four-layer CNN.
In addition, some works made innovations in network design. The authors of [122] proposed to use dilated convolutions and share parameters in the cascade networks. Dense connections were added to an unrolling network in [123]. The authors of [124] proposed to use a cascade network for k-space reconstruction followed by a network for image.
Recent works considered other types of loss function. The authors of [125] considered to use adversarial loss function and proposed a trick to balance different loss functions. In [126] perceptual loss was used. In addition, attention layers were applied to U-Net as the P image . In [127], three cascade networks are connected sequentially and their output is concatenated to the last convolutional layer. In each cascade network, convolutions with different strides are used to utilized different scale information. Each network is a RNN which is equivalent to an unrolling form network. In [128], multi-contrast MRI reconstruction was considered and a convolution-shared network was proposed.

4.3.2. Parallel Imaging

The authors of [129] utilized a U-Net to reconstruct parallel imaging. WGAN [130] was exploited in [131] and three sequentially connected U-Nets were used as the generator. In training stage, MSE loss of image and k-space, adversarial loss and perceptual loss are applied. The authors of [132] considered 3d MRI reconstruction. The proposed network contains two parts, MS-net for feature extraction and RC-net for reconstruction.
As for methods of unrolling form, most works applied similar network architectures in Section 4.3.1 and the main difference is on the P dc since each coil has its own data consistency equation. The authors of [133] unrolled a proximal gradient algorithm to a network and applied it to 3d MRI reconstruction. CNNs are used to replace proximal operators. A U-Net was used in [134] to substitute the proximal operator in the ADMM algorithm. In [135], CNNs and data consistency layers are connected alternately and two different process of multi-coil were considered. The authors of [136] proposed to unrolled a first-order gradient method and the regularization term in object function is related to a neural network. The authors of [137] applied the method in [43] to parallel MR imaging. The authors of [138] proposed to utilized variable splitting algorithm. The object function has the following form:
min m , u , x i λ 2 i = 1 n c D F x i y i 2 2 + R ( u ) + α 2 i = 1 n c x i S i m 2 2 + β 2 u m 2 2 ,
where D is the sampling matrix and m is the reconstructed image. x i represents the the result of image multiplying by the sensitive matrix of the ith coil. A denoiser network is used to replace the computation of arg min u β 2 u m k 2 2 + R ( u ) . In [139] complex number operation was combined with neural networks. The authors of [140] proposed to jointly estimate images and sensitive matrices in a unrolling network. The original object function is as follows:
1 2 l M F V l y l 2 2 + ρ 2 l S l U V l 2 2 + β l R S l + λ P ( U ) ,
where U represents the image and S l is the sensitive matrix. The proximal operator of R ( S l ) and P ( U ) of corresponding iteration are substituted by two sub-blocks and each sub-block contains several sub-networks.
In order to guarantee the convergence of the unrolling network, ref. [141] proposed to use a judgement condition to decide whether to receive the result of the neural network P image .
Besides the reviewed works above, deep learning methods consisting of single P image can be seen in [92,142,143,144,145,146,147,148,149,150,151,152,153,154,155,156,157], etc. Other unrolling form deep learning methods can be seen in [158,159,160,161,162,163,164,165,166,167,168,169,170], etc.

4.4. Latent Variable Search of Generative Models

A recent work [171] belongs to this category, the purpose of which is to reconstruct parallel imaging of MRI. First, a GAN is trained to generate MRI images. When reconstructing images, besides latent variable, parameters of generative network is also optimized. The reconstruction process contains two stage. In the first stage, the following optimization problem is solved:
min z R d 1 2 A G θ ( z ) y 2 , s . t . z d .
Then, in the second stage, latent variable and parameters are both optimized:
min ( z , θ ) R d × R l 1 2 A G θ ( z ) y 2 , s . t . z d .
In addition, the sensitive matrices are estimated by the ESPIRiT algorithm.

4.5. Neural Networks Based Probability Models

The authors of [172] proposed to estimate the prior distribution by a trained VAE [173] and use it to optimize the Bayesian model through projection onto convex sets (POCS) algorithm. PixelCNN++ [174] was used in [175] to represent the prior distribution and a gradient-projection algorithm was applied to solved the Bayesian model.

4.6. Unsupervised Methods

In this category, most related works are based on the DIP method which has been reviewed in Section 2.6. The authors of [176] used DIP directly and made no modification. The authors of [177] applied it to dynamic MRI reconstruction and exploited linear interpolation to obtain inputs of the network for continuous multi-frame images. Meanwhile, the authors of [178] used measurements and zero-fill reconstruction as labels to train a network. The loss function is similar to the one in DIP method and has the following form:
L ( y , y ^ ) = α y y ^ 1 + β Φ y S Φ ( x ^ ) 1 + γ I θ ( y ) I θ ( y ^ ) 1 .
Here, y is not measurement but zero-fill reconstruction from under-sampled k-space coefficients and y ^ is the reconstructed image. They are used as the input of the network I θ . x ^ is the output of I θ ; S is the sampling matrix and Φ is Fourier transform. Besides the DIP method, in [179] a novel loss function was designed to implement an unsupervised training scheme. The under-sampled k-space index is divided into two groups which is denoted by Ω = Θ Λ . Correspondingly, measurement y and measuring matrix can also be divided into ( y Θ , E Θ ) and ( y Λ , E Λ ) . ( y Λ , E Λ ) is used as labels and ( y Θ , E Θ ) as training inputs. Then the loss function has the following form:
1 N i = 1 N L y Λ i , E Λ i f y Θ i , E Θ i ; θ ,
where f represents the reconstruction algorithm. Though the method of [179] belongs to unsupervised learning, f in this work is an unrolling form network which similar to the ones reviewed in Section 4.3.

4.7. Discussion

Different from CT reconstruction, the measurement matrix in MRI reconstruction is very simple (Fourier transform) and computable. Thus, it is more easy to propose various categories of methods for reconstruction similar to CS reconstruction. We observed that reviewed works cover all the categories mentioned in Section 2 and many unrolling methods were proposed. The trend of deep learning method is similar to CS reconstruction. However, the peculiarity of kspace data and parallel imaging distinguish MRI reconstruction from other medical image reconstruction. The sensitive matrics in parallel imaging also present a challenge for researchers.

5. Deep Learning Methods for Positron-Emission Tomography

5.1. Overview

Positron-emission Tomography is another common medical imaging tool which utilizes radioactive material. It needs some detectors to receive photons emitted by radioactive element. To reduce the risk, the reconstruction from low-dose PET is desired. Different to CT, the number of detectors is not decreased in most low-dose PET. Thus, in low-dose PET reconstruction, the forward model somehow cannot be deemed as a compressed sensing problem. However, many works that focus on the reconstruction of PET employ deep learning methods similar to CS reconstruction. Those methods can also be classified as some categories discussed in Section 2. Therefore, in this section, we still review some related works of low-dose PET reconstruction.

5.2. Neural Networks as Image Projections

In [180], the initial reconstruction results are obtained by a traditional method with different weights. Then the patches of those results are fed into a fully-connected network to produce better reconstruction. The authors of [181] proposed to use a U-Net to transform low-dose PET images to the ones of high quality. The global residual learning structure is utilized and L1 norm is used as the loss function. In addition, several adjacent slices are the input of network. The perceptual loss was exploited in [182]. The network is trained by simulated data at first and then refined by real data. The authors of [183] proposed to use a conditional WGAN to perform 3d reconstruction. The backbone of the generative network is a U-Net and the input is a 3d low-dose PET image. At the beginning and final part of the generative network, 3d convolutions are used while in the middle part 2d convolutions are applied. The training scheme includes two stages. In the first stage, MSE and SSIM loss are used to train the generative network, and in the second stage, adversarial loss and perceptual loss are added to train the model. Similar to [183], in [184] a conditional GAN was proposed. The generative network is a 3d U-Net and its input is patches of low-dose PET images. Besides, a multi-GAN refinement treatment was proposed for better performance. The output of former GAN is used as the input of the next one and each GAN is trained one by one. Some works attempted to exploit other modality information to help PET reconstruction. In [185,186], MR images are fed into network as extra input. For PET, the measurement is also called sinogram since it is similar to CT. In some works it is considered to be the input of networks rather than the initial reconstruction. The authors of [187] used sinogram as the input of a conditional GAN whose generative network is still a U-Net. The authors of [188] used it as the input of a CNN and preprocessed the sinogram before feeding it to the network to reduce the effect of random noise.
Some works are proposed to utilize multiple networks or unroll an iterative algorithm. The authors of [189] proposed a Learned Primal-Dual method which contains two types of U-Nets: one for images and the other for sinogram. These U-Nets are ordered alternatively and connected to each other through measuring matrix and its transposition. The authors of [190] proposed to combine a denoising network and an iterative algorithm. A denoising model DnCNN is trained in advance and added to the logarithmic likelihood function as a regularization. The object function has the following form:
i = 1 N m [ A x ] i + r i y i log [ A x ] i + r i + β 2 x q f w ( x ) b 2 2 .
Equation (56) is solved by ADMM after variable splitting. THhe authors of [191] proposed MAPEM-Net which unrolls the ADMM algorithm. A U-Net is used to replace the proximal operator of regularization term. The authors of [192] proposed to use a trained conditional GAN as a constraint for the logarithmic likelihood function. The optimization problem is as follows:
max x , α η L y | P α + s + r ) ) + L ( y | P x + s + r ) , s . t . , x = f ( α ) ,
where f represents the generative network and α is five slices of low-dose images. ADMM algorithm is applied to solve it.

5.3. Latent Variable Search of Generative Models

The method proposed by [193] can be regarded to belong this category. However, there is something different. In [193], the generative model is a denoising U-Net instead of a GAN or VAE. After pre-training the network, it is used as a constraint for logarithmic likelihood function. The optimization problem can be written as follows:
max x L ( y | x ) ,
s . t . x = f ( α ) ,
where L represents the likelihood function, f is the network and α is the input. It is solved by ADMM algorithm.

5.4. Unsupervised Methods

In this part, most works also applied DIP method to PET reconstruction. Similar to [193], in [194], a logarithmic likelihood function with a constraint that x is the output of a network is the object function for optimization. However, the network is untrained and the input is fixed. Then it turns to be a DIP-like problem. L-BFGS algorithm is used to solve it. The authors of [195] applied an almost same framework for PET reconstruction except that the input of network is replace by a related CT or MR image. The authors of [196] proposed to combine DIP and non-negative matrix factorization to reconstruct dynamic PET images. DIP is used for image representation and non-negative matrix factorization for controlling temporal sparsity. The object function has the following form:
minimize Θ , B L : = D KL Y P A B T + α A T p , 2 2 + β B QV 2 ,
s . t . A = a 1 , , a R 0 , B 0 ,
a r = ϕ u | θ r [ 0 , 1 ] N i ,
a r = 1 for r = 1 , 2 , , R .
For other works using DIP unsupervised method, the reader can refer to [197,198,199].

5.5. Discussion

Because the measuring model of PET is the most complex (Poisson noise and ill-posed measurement matrix), it is hard to design P dc and different unrolling methods. Therefore, image domain is usually considered in PET reconstruction and it is often regarded as a denoising problem using popular U-Net, which is somehow similar to CT reconstruction.

6. Discussion and Future Directions

We have reviewed many works of the deep learning application in CS, CT, MRI and PET reconstruction. Though they are different in details, these works hold a common character, satisfying framework F which is described in Section 1. In general, most neural networks play the role of P image . Therefore, the reconstruction framework is the same to traditional methods.
We may ask: what is the advantages of deep learning? In the framework F , P image is the key part, because in most cases, the P dc is easy to derive. Thus, the performance of a reconstruction algorithm usually depends on the design of P image . For traditional methods, P image is derived from a hand-designed image prior distribution or regularization terms. Even if some method are similar to data-driven methods such as dictionary learning, the L1 norm and the form of linear transform are determined in advance. The drawback of the hand-designed model is that it may be insufficient or inaccurate to depict the real prior distribution of signals. However, deep learning holds two advantages that make it successful. Firstly, it is data-driven. If a large dataset is available, the model can directly utilize the distribution information hidden in training data. Secondly, it allow researchers to design more complex and flexible model to better represent the image prior distribution.
However, deep learning also has a disadvantage that has not been solved well. In [200], three tests were used to inspect the stability of deep learning. Several popular models are compared to traditional methods. The results show that there is some instability problem in deep learning, while there isstrong stability for traditional methods. The lack of training data may be another handicap for medical applications.
Nevertheless, deep learning provides a powerful tool which can be used to learn prior information from data. More specifically, there are three types of models proposed in current research. Neural networks are used to depict P image , p ( x ) or M image directly. Usually, a CNN denoising-like model is used to represent P image , the projection operator (see Section 2.3, Section 3.3, Section 4.3 and Section 5.2). In the MAP method, the network is exploited to compute the prior distribution p ( x ) (see Section 2.5 and Section 4.5). Generative models are utilized to depict image manifold M image (see Section 2.4, Section 4.4, Section 4.6, Section 5.3 and Section 5.4).
As for future research, how to design more efficient network is a clear direction. We have seen three types of deep learning model. Which one is the best? What is the relationship between different models, P image , M image and p ( x ) . These questions have not been answered. Actually, P image , M image and p ( x ) are different facets of one thing, the image prior. Combining them all may a feasible way to design novel networks. The breakthrough of deep learning theory ought to be helpful. It may tell us how a network plays the role of P image or how it can represent a complex manifold. It can also provide new ideas to train the model. In addition, the statistical properties of image or signal prior distribution can inspire researchers to design more feasible and robust network architectures. For example, the property of multi-scale has been considered in many works. Few shot learning, robustness of networks and computation efficiency are also worthy of attention. Besides, the theoretical properties of deep learning reconstruction methods, including the existence and uniqueness of solution and the convergence of algorithm, are important research directions in the future. Another important issue discussed less in this review is the measurement noise and artifacts, which will lead to noisy images in real life. It is necessary to alleviate the effect of noise and artifacts. One of methods is pre–Processing of these noisy images, for example, denoising using 1st and 2nd generation wavelets [201]. Readers can refer to [202] for more works about it. Security [203] and privacy-perserving problem [204] are also important in image reconstruction, especially in medical image reconstruction tasks. However, they have not been studied deeply. In addition, compressed sensing, or inverse problem also exists in the area of surveillance [205], medical [206], agriculture [207], speech [208] and telecommunications. Our proposed framework may be helpful to inspire researchers to improve their works.

7. Conclusions

Deep learning has been proved to be successful in CS reconstruction. In this paper, we review some works on it and its medical applications using deep learning methods. A framework F is derived to better understand these approaches. We define two projection operators toward image prior and data consistency, respectively, and any reconstruction algorithm can be decomposed to the two parts. Based on it, several categories are analyzed and relationship between them is built. It also helps us to connect deep learning methods to traditional iterative algorithms. Our analysis illustrates that the key to solve CS problem and its medical applications is how to depict the image prior to this. We hope that the proposed framework and our observation may provide a new perspective to improve the current work.

Author Contributions

Conceptualization, Y.X.; investigation, Y.X.; writing—original draft preparation, Y.X.; writing—review and editing, Q.L.; visualization, Y.X.; supervision, Q.L.; project administration, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liang, D.; Cheng, J.; Ke, Z.; Ying, L. Deep mri reconstruction: Unrolled optimization algorithms meet neural networks. arXiv 2019, arXiv:1907.11711. [Google Scholar]
  2. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  3. Gregor, K.; LeCun, Y. Learning fast approximations of sparse coding. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 399–406. [Google Scholar]
  4. Kamilov, U.S.; Mansour, H. Learning optimal nonlinearities for iterative thresholding algorithms. IEEE Signal Process. Lett. 2016, 23, 747–751. [Google Scholar] [CrossRef] [Green Version]
  5. Bostan, E.; Kamilov, U.S.; Waller, L. Learning-based image reconstruction via parallel proximal algorithm. IEEE Signal Process. Lett. 2018, 25, 989–993. [Google Scholar] [CrossRef] [Green Version]
  6. Mahapatra, D.; Mukherjee, S.; Seelamantula, C.S. Deep sparse coding using optimized linear expansion of thresholds. arXiv 2017, arXiv:1705.07290. [Google Scholar]
  7. Zhang, J.; Ghanem, B. ISTA-Net: Interpretable optimization-inspired deep network for image compressive sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 1828–1837. [Google Scholar]
  8. Mukherjee, S.; Mahapatra, D.; Seelamantula, C.S. DNNs for sparse coding and dictionary learning. In Proceedings of the NIPS Bayesian Deep Learning Workshop, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  9. Lu, X.; Dong, W.; Wang, P.; Shi, G.; Xie, X. Convcsnet: A convolutional compressive sensing framework based on deep learning. arXiv 2018, arXiv:1801.10342. [Google Scholar]
  10. Pokala, P.K.; Mahurkar, A.G.; Seelamantula, C.S. FirmNet: A Sparsity Amplified Deep Network for Solving Linear Inverse Problems. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2982–2986. [Google Scholar]
  11. Perdios, D.; Besson, A.; Rossinelli, P.; Thiran, J.P. Learning the weight matrix for sparsity averaging in compressive imaging. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3056–3060. [Google Scholar]
  12. Zhang, X.; Yuan, X.; Carin, L. Nonlocal low-rank tensor factor analysis for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 8232–8241. [Google Scholar]
  13. Mousavi, A.; Patel, A.B.; Baraniuk, R.G. A deep learning approach to structured signal recovery. In Proceedings of the 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton), Champaign, IL, USA, 30 September–2 October 2015; pp. 1336–1343. [Google Scholar]
  14. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 449–458. [Google Scholar]
  15. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  16. Mousavi, A.; Baraniuk, R.G. Learning to invert: Signal recovery via deep convolutional networks. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2272–2276. [Google Scholar]
  17. Shi, W.; Jiang, F.; Zhang, S.; Zhao, D. Deep networks for compressed image sensing. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 877–882. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  19. Yao, H.; Dai, F.; Zhang, S.; Zhang, Y.; Tian, Q.; Xu, C. Dr2-net: Deep residual reconstruction network for image compressive sensing. Neurocomputing 2019, 359, 483–493. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, Y.; Bai, H.; Zhao, L.; Zhao, Y. Cascaded reconstruction network for compressive image sensing. EURASIP J. Image Video Process. 2018, 2018, 77. [Google Scholar] [CrossRef]
  21. Huang, H.; Nie, G.; Zheng, Y.; Fu, Y. Image restoration from patch-based compressed sensing measurement. Neurocomputing 2019, 340, 145–157. [Google Scholar] [CrossRef]
  22. Xie, X.; Wang, C.; Du, J.; Shi, G. Full image recover for block-based compressive sensing. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 1–6. [Google Scholar]
  23. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  24. Lohit, S.; Kulkarni, K.; Kerviche, R.; Turaga, P.; Ashok, A. Convolutional neural networks for noniterative reconstruction of compressively sensed images. IEEE Trans. Comput. Imaging 2018, 4, 326–340. [Google Scholar] [CrossRef] [Green Version]
  25. Du, J.; Xie, X.; Wang, C.; Shi, G. Perceptual compressive sensing. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Guangzhou, China, 23–26 November 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 268–279. [Google Scholar]
  26. Zur, Y.; Adler, A. Deep Learning of Compressed Sensing Operators with Structural Similarity Loss. arXiv 2019, arXiv:1906.10411. [Google Scholar]
  27. Zhang, Z.; Gao, D.; Xie, X.; Shi, G. Dual-Channel Reconstruction Network for Image Compressive Sensing. Sensors 2019, 19, 2549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Xu, K.; Zhang, Z.; Ren, F. Lapran: A scalable laplacian pyramid reconstructive adversarial network for flexible compressive sensing reconstruction. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 485–500. [Google Scholar]
  29. Shi, W.; Jiang, F.; Liu, S.; Zhao, D. Scalable Convolutional Neural Network for Image Compressed Sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 12290–12299. [Google Scholar]
  30. Chen, D.; Davies, M.E. Deep Decomposition Learning for Inverse Imaging Problems. arXiv 2019, arXiv:1911.11028. [Google Scholar]
  31. Li, W.; Liu, F.; Jiao, L.; Hu, F. Multi-Scale Residual Reconstruction Neural Network with Non-Local Constraint. IEEE Access 2019, 7, 70910–70918. [Google Scholar] [CrossRef]
  32. Rick Chang, J.; Li, C.L.; Poczos, B.; Vijaya Kumar, B.; Sankaranarayanan, A.C. One Network to Solve Them All–Solving Linear Inverse Problems Using Deep Projection Models. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5888–5897. [Google Scholar]
  33. Zhao, C.; Zhang, J.; Wang, R.; Gao, W. CREAM: CNN-REgularized ADMM framework for compressive-sensed image reconstruction. IEEE Access 2018, 6, 76838–76853. [Google Scholar] [CrossRef]
  34. Kelly, B.; Matthews, T.P.; Anastasio, M.A. Deep learning-guided image reconstruction from incomplete data. arXiv 2017, arXiv:1709.00584. [Google Scholar]
  35. Metzler, C.; Mousavi, A.; Baraniuk, R. Learned D-AMP: Principled neural network based compressive image recovery. Adv. Neural Inf. Process. Syst. 2017, 30, 1772–1783. [Google Scholar]
  36. Zhou, S.; He, Y.; Liu, Y.; Li, C. Multi-Channel Deep Networks for Block-Based Image Compressive Sensing. arXiv 2019, arXiv:1908.11221. [Google Scholar] [CrossRef]
  37. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Aggarwal, H.K.; Mani, M.P.; Jacob, M. Modl: Model-based deep learning architecture for inverse problems. IEEE Trans. Med. Imaging 2018, 38, 394–405. [Google Scholar] [CrossRef] [PubMed]
  39. Diamond, S.; Sitzmann, V.; Heide, F.; Wetzstein, G. Unrolled optimization with deep priors. arXiv 2017, arXiv:1705.08041. [Google Scholar]
  40. Adler, J.; Öktem, O. Learned primal-dual reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1322–1332. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Zhang, Q.; Chen, Y. Extra Proximal-Gradient Inspired Non-local Network. arXiv 2019, arXiv:1911.07144. [Google Scholar]
  42. Raj, A.; Li, Y.; Bresler, Y. GAN-Based Projector for Faster Recovery with Convergence Guarantees in Linear Inverse Problems. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 5602–5611. [Google Scholar]
  43. Zhang, M.; Yuan, Y.; Zhang, F.; Wang, S.; Wang, S.; Liu, Q. Multi-Noise and Multi-Channel Derived Prior Information for Grayscale Image Restoration. IEEE Access 2019, 7, 150082–150092. [Google Scholar] [CrossRef]
  44. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 19–21 June 2018; pp. 773–782. [Google Scholar]
  45. He, Z.; Zhou, J.; Liang, D.; Wang, Y.; Liu, Q. Learning Priors in High-frequency Domain for Inverse Imaging Reconstruction. arXiv 2019, arXiv:1910.11148. [Google Scholar]
  46. Bora, A.; Jalal, A.; Price, E.; Dimakis, A.G. Compressed sensing using generative models. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 537–546. [Google Scholar]
  47. Dhar, M.; Grover, A.; Ermon, S. Modeling sparse deviations for compressed sensing using generative models. arXiv 2018, arXiv:1807.01442. [Google Scholar]
  48. Kabkab, M.; Samangouei, P.; Chellappa, R. Task-aware compressed sensing with generative adversarial networks. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  49. Chen, L.; Yang, H. Generative Imaging and Image Processing via Generative Encoder. arXiv 2019, arXiv:1905.13300. [Google Scholar]
  50. Xu, S.; Zeng, S.; Romberg, J. Fast Compressive Sensing Recovery Using Generative Models with Structured Latent Variables. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2967–2971. [Google Scholar]
  51. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2180–2188. [Google Scholar]
  52. Wu, Y.; Rosca, M.; Lillicrap, T. Deep compressed sensing. arXiv 2019, arXiv:1905.06723. [Google Scholar]
  53. Dave, A.; Kumar, A.; Mitra, K. Compressive image recovery using recurrent generative model. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1702–1706. [Google Scholar]
  54. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  55. Dave, A.; Vadathya, A.K.; Subramanyam, R.; Baburajan, R.; Mitra, K. Solving inverse computational imaging problems using deep pixel-level prior. IEEE Trans. Comput. Imaging 2018, 5, 37–51. [Google Scholar] [CrossRef] [Green Version]
  56. Van den Oord, A.; Kalchbrenner, N.; Vinyals, O.; Espeholt, L.; Graves, A.; Kavukcuoglu, K. Conditional image generation with pixelcnn decoders. arXiv 2016, arXiv:1606.05328. [Google Scholar]
  57. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Deep image prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 9446–9454. [Google Scholar]
  58. Van Veen, D.; Jalal, A.; Soltanolkotabi, M.; Price, E.; Vishwanath, S.; Dimakis, A.G. Compressed sensing with deep image prior and learned regularization. arXiv 2018, arXiv:1806.06438. [Google Scholar]
  59. Ravula, S.; Dimakis, A.G. One-dimensional deep image prior for time series inverse problems. arXiv 2019, arXiv:1904.08594. [Google Scholar]
  60. Leong, O.; Sakla, W. Low Shot Learning with Untrained Neural Networks for Imaging Inverse Problems. arXiv 2019, arXiv:1910.10797. [Google Scholar]
  61. Kobler, E.; Muckley, M.; Chen, B.; Knoll, F.; Hammernik, K.; Pock, T.; Sodickson, D.; Otazo, R. Variational deep learning for low-dose computed tomography. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6687–6691. [Google Scholar]
  62. Zhang, H.; Dong, B.; Liu, B. JSR-Net: A deep network for joint spatial-radon domain CT reconstruction from incomplete data. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3657–3661. [Google Scholar]
  63. Boublil, D.; Zibulevsky, M.; Elad, M. Compressed Sensing and Computed Tomography with Deep Neural Networks. 2015. Available online: https://pdfs.semanticscholar.org/c1de/cfd99ce8affed9fef1ae9292a0f242493813.pdf (accessed on 11 January 2022).
  64. Zhao, J.; Chen, Z.; Zhang, L.; Jin, X. Few-view CT reconstruction method based on deep learning. In Proceedings of the 2016 IEEE Nuclear Science Symposium, Medical Imaging Conference and Room-Temperature Semiconductor Detector Workshop (NSS/MIC/RTSD), Strasbourg, France, 29 October–6 November 2016; pp. 1–4. [Google Scholar]
  65. Zhang, H.; Li, L.; Qiao, K.; Wang, L.; Yan, B.; Li, L.; Hu, G. Image prediction for limited-angle tomography via deep learning with convolutional neural network. arXiv 2016, arXiv:1607.08707. [Google Scholar]
  66. Jin, K.H.; McCann, M.T.; Unser, M. BPConvNet for Compressed Sensing Recovery in Bioimaging. 2016. Available online: http://spars2017.lx.it.pt/index_files/papers/SPARS2017_Paper_119.pdf (accessed on 11 January 2022).
  67. Han, Y.S.; Yoo, J.; Ye, J.C. Deep residual learning for compressed sensing CT reconstruction via persistent homology analysis. arXiv 2016, arXiv:1611.06391. [Google Scholar]
  68. Han, Y.; Ye, J.C. Framing U-Net via deep convolutional framelets: Application to sparse-view CT. IEEE Trans. Med. Imaging 2018, 37, 1418–1429. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. Yang, Q.; Yan, P.; Kalra, M.K.; Wang, G. CT image denoising with perceptive deep neural networks. arXiv 2017, arXiv:1702.07019. [Google Scholar]
  70. Choi, K.; Kim, S.W.; Lim, J.S. Real-time image reconstruction for low-dose CT using deep convolutional generative adversarial networks (GANs). In Medical Imaging 2018: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2018; Volume 10573, p. 1057332. [Google Scholar]
  71. Gu, J.; Ye, J.C. Multi-scale wavelet domain residual learning for limited-angle CT reconstruction. arXiv 2017, arXiv:1703.01382. [Google Scholar]
  72. Ye, D.H.; Buzzard, G.T.; Ruby, M.; Bouman, C.A. Deep back projection for sparse-view CT reconstruction. In Proceedings of the 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Anaheim, CA, USA, 26–29 November 2018; pp. 1–5. [Google Scholar]
  73. Dong, X.; Vekhande, S.; Cao, G. Sinogram interpolation for sparse-view micro-CT with deep learning neural network. In Medical Imaging 2019: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10948, p. 109482O. [Google Scholar]
  74. Li, Z.; Cai, A.; Wang, L.; Zhang, W.; Tang, C.; Li, L.; Liang, N.; Yan, B. Promising Generative Adversarial Network Based Sinogram Inpainting Method for Ultra-Limited-Angle Computed Tomography Imaging. Sensors 2019, 19, 3941. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  75. Würfl, T.; Ghesu, F.C.; Christlein, V.; Maier, A. Deep learning computed tomography. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 432–440. [Google Scholar]
  76. Chen, H.; Zhang, Y.; Zhou, J.; Wang, G. Deep learning for low-dose CT. In Developments in X-ray Tomography XI; International Society for Optics and Photonics: Bellingham, WA, USA, 2017; Volume 10391, p. 103910I. [Google Scholar]
  77. Nguyen, T.C.; Bui, V.; Nehmetallah, G. Computational optical tomography using 3-D deep convolutional neural networks. Opt. Eng. 2018, 57, 043111. [Google Scholar]
  78. Clark, D.; Badea, C. Convolutional regularization methods for 4D, X-ray CT reconstruction. In Medical Imaging 2019: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10948, p. 109482A. [Google Scholar]
  79. Liu, J.; Zhang, Y.; Zhao, Q.; Lv, T.; Wu, W.; Cai, N.; Quan, G.; Yang, W.; Chen, Y.; Luo, L.; et al. Deep iterative reconstruction estimation (DIRE): Approximate iterative reconstruction estimation for low dose CT imaging. Phys. Med. Biol. 2019, 64, 135007. [Google Scholar] [CrossRef] [PubMed]
  80. Cong, W.; Shan, H.; Zhang, X.; Liu, S.; Ning, R.; Wang, G. Deep-learning-based breast CT for radiation dose reduction. In Developments in X-ray Tomography XII; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11113, p. 111131L. [Google Scholar]
  81. Beaudry, J.; Esquinas, P.L.; Shieh, C.C. Learning from our neighbours: A novel approach on sinogram completion using bin-sharing and deep learning to reconstruct high quality 4DCBCT. In Medical Imaging 2019: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10948, p. 1094847. [Google Scholar]
  82. Chen, H.; Zhang, Y.; Chen, Y.; Zhang, J.; Zhang, W.; Sun, H.; Lv, Y.; Liao, P.; Zhou, J.; Wang, G. LEARN: Learned experts’ assessment-based reconstruction network for sparse-data CT. IEEE Trans. Med. Imaging 2018, 37, 1333–1347. [Google Scholar] [CrossRef]
  83. He, J.; Yang, Y.; Wang, Y.; Zeng, D.; Bian, Z.; Zhang, H.; Sun, J.; Xu, Z.; Ma, J. Optimizing a parameterized plug-and-play ADMM for iterative low-dose CT reconstruction. IEEE Trans. Med. Imaging 2018, 38, 371–382. [Google Scholar] [CrossRef]
  84. Wang, J.; Zeng, L.; Wang, C.; Guo, Y. ADMM-based deep reconstruction for limited-angle CT. Phys. Med. Biol. 2019, 64, 115011. [Google Scholar] [CrossRef]
  85. Cheng, W.; Wang, Y.; Chi, Y.; Xie, X.; Duan, Y. Learned Full-Sampling Reconstruction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 375–384. [Google Scholar]
  86. Chun, I.Y.; Zheng, X.; Long, Y.; Fessler, J.A. BCD-Net for Low-dose CT Reconstruction: Acceleration, Convergence, and Generalization. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 31–40. [Google Scholar]
  87. Ding, Q.; Chen, G.; Zhang, X.; Huang, Q.; Gao, H.J.H. Low-Dose CT with Deep Learning Regularization via Proximal Forward Backward Splitting. arXiv 2019, arXiv:1909.09773. [Google Scholar] [CrossRef] [Green Version]
  88. Hauptmann, A.; Adler, J.; Arridge, S.; Öktem, O. Multi-Scale Learned Iterative Reconstruction. arXiv 2019, arXiv:1908.00936. [Google Scholar] [CrossRef]
  89. Li, Z.; Ye, S.; Long, Y.; Ravishankar, S. SUPER Learning: A Supervised-Unsupervised Framework for Low-Dose CT Image Reconstruction. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  90. He, J.; Wang, Y.; Yang, Y.; Bian, Z.; Zeng, D.; Sun, J.; Xu, Z.; Ma, J. LdCT-net: Low-dose CT image reconstruction strategy driven by a deep dual network. In Medical Imaging 2018: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2018; Volume 10573, p. 105733G. [Google Scholar]
  91. Wu, D.; Kim, K.; Li, Q. Computationally efficient deep neural network for computed tomography image reconstruction. Med. Phys. 2019, 46, 4763–4776. [Google Scholar] [CrossRef] [Green Version]
  92. Dedmari, M.A.; Conjeti, S.; Estrada, S.; Ehses, P.; Stöcker, T.; Reuter, M. Complex fully convolutional neural networks for mr image reconstruction. In International Workshop on Machine Learning for Medical Image Reconstruction; Springer: Berlin/Heidelberg, Germany, 2018; pp. 30–38. [Google Scholar]
  93. Pruessmann, K.P.; Weiger, M.; Scheidegger, M.B.; Boesiger, P. SENSE: Sensitivity encoding for fast MRI. Magn. Reson. Med. Off. J. Int. Soc. Magn. Reson. Med. 1999, 42, 952–962. [Google Scholar] [CrossRef]
  94. Sun, J.; Li, H.; Xu, Z. Deep ADMM-Net for compressive sensing MRI. Adv. Neural Inf. Process. Syst. 2016, 29, 10–18. [Google Scholar]
  95. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A Deep Learning Approach for Image Compressive Sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 521–538. [Google Scholar] [CrossRef] [PubMed]
  96. Li, Y.; Huang, L.; Yin, Y.; Wang, Y.; Gui, G. ADMM-Net for Robust Compressive Sensing Image Reconstruction in the Presence of Symmetric α-Stable Noise. In Proceedings of the APSIPA Annual Summit and Conference, Hawaii, HI, USA, 12–15 November 2018; Volume 2018, pp. 12–15. [Google Scholar]
  97. Liu, Y.; Liu, Q.; Zhang, M.; Yang, Q.; Wang, S.; Liang, D. IFR-Net: Iterative Feature Refinement Network for Compressed Sensing MRI. IEEE Trans. Comput. Imaging 2019, 6, 434–446. [Google Scholar] [CrossRef] [Green Version]
  98. Hammernik, K.; Klatzer, T.; Kobler, E.; Recht, M.P.; Sodickson, D.K.; Pock, T.; Knoll, F. Learning a variational network for reconstruction of accelerated MRI data. Magn. Reson. Med. 2018, 79, 3055–3071. [Google Scholar] [CrossRef] [PubMed]
  99. Uecker, M.; Lai, P.; Murphy, M.J.; Virtue, P.; Elad, M.; Pauly, J.M.; Vasanawala, S.S.; Lustig, M. ESPIRiT—An eigenvalue approach to autocalibrating parallel MRI: Where SENSE meets GRAPPA. Magn. Reson. Med. 2014, 71, 990–1001. [Google Scholar] [CrossRef] [Green Version]
  100. Chen, Y.; Xiao, T.; Li, C.; Liu, Q.; Wang, S. Model-based Convolutional De-Aliasing Network Learning for Parallel MR Imaging. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 30–38. [Google Scholar]
  101. Chen, F.; Taviani, V.; Malkiel, I.; Cheng, J.Y.; Tamir, J.I.; Shaikh, J.; Chang, S.T.; Hardy, C.J.; Pauly, J.M.; Vasanawala, S.S. Variable-density single-shot fast spin-echo MRI with deep learning reconstruction by using variational networks. Radiology 2018, 289, 366–373. [Google Scholar] [CrossRef] [Green Version]
  102. Ravishankar, S.; Lahiri, A.; Blocker, C.; Fessler, J.A. Deep dictionary-transform learning for image reconstruction. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, WA, USA, 4–7 April 2018; pp. 1208–1212. [Google Scholar]
  103. Lu, T.; Zhang, X.; Huang, Y.; Yang, Y.; Guo, G.; Bao, L.; Huang, F.; Guo, D.; Qu, X. pISTA-SENSE-ResNet for Parallel MRI Reconstruction. arXiv 2019, arXiv:1910.00650. [Google Scholar]
  104. Wang, S.; Su, Z.; Ying, L.; Peng, X.; Zhu, S.; Liang, F.; Feng, D.; Liang, D. Accelerating magnetic resonance imaging via deep learning. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 514–517. [Google Scholar]
  105. Lee, D.; Yoo, J.; Ye, J.C. Deep artifact learning for compressed sensing and parallel MRI. arXiv 2017, arXiv:1703.01120. [Google Scholar]
  106. Yu, S.; Dong, H.; Yang, G.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Firmin, D.; et al. Deep de-aliasing for fast compressive sensing MRI. arXiv 2017, arXiv:1705.07137. [Google Scholar]
  107. Sandino, C.M.; Dixit, N.; Cheng, J.Y.; Vasanawala, S.S. Deep Convolutional Neural Networks for Accelerated Dynamic Magnetic Resonance Imaging. 2017. Available online: http://cs231n.stanford.edu/reports/2017/pdfs/513.pdf (accessed on 11 January 2022).
  108. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Guo, Y.; et al. DAGAN: Deep de-aliasing generative adversarial networks for fast compressed sensing MRI reconstruction. IEEE Trans. Med. Imaging 2017, 37, 1310–1321. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  109. Deora, P.; Vasudeva, B.; Bhattacharya, S.; Pradhan, P.M. Robust Compressive Sensing MRI Reconstruction using Generative Adversarial Networks. arXiv 2019, arXiv:1910.06067. [Google Scholar]
  110. Mardani, M.; Gong, E.; Cheng, J.Y.; Vasanawala, S.S.; Zaharchuk, G.; Xing, L.; Pauly, J.M. Deep generative adversarial neural networks for compressive sensing MRI. IEEE Trans. Med. Imaging 2018, 38, 167–179. [Google Scholar] [CrossRef] [PubMed]
  111. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul Smolley, S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  112. Li, Z.; Zhang, T.; Wan, P.; Zhang, D. SEGAN: Structure-enhanced generative adversarial network for compressed sensing MRI reconstruction. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 1012–1019. [Google Scholar]
  113. Xu, C.; Tao, J.; Ye, Z.; Xu, J.; Kainat, W. Adversarial training and dilated convolutions for compressed sensing MRI. In Proceedings of the Eleventh International Conference on Digital Image Processing (ICDIP 2019), Guangzhou, China, 10–13 May 2019; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11179, p. 111793T. [Google Scholar]
  114. Quan, T.M.; Nguyen-Duc, T.; Jeong, W.K. Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss. IEEE Trans. Med. Imaging 2018, 37, 1488–1497. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  115. Hyun, C.M.; Kim, H.P.; Lee, S.M.; Lee, S.; Seo, J.K. Deep learning for undersampled MRI reconstruction. Phys. Med. Biol. 2018, 63, 135007. [Google Scholar] [CrossRef]
  116. Souza, R.; Frayne, R. A hybrid frequency-domain/image-domain deep network for magnetic resonance image reconstruction. In Proceedings of the 2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Rio de Janeiro, Brazil, 28–30 October 2019; pp. 257–264. [Google Scholar]
  117. Eo, T.; Jun, Y.; Kim, T.; Jang, J.; Lee, H.J.; Hwang, D. KIKI-net: Cross-domain convolutional neural networks for reconstructing undersampled magnetic resonance images. Magn. Reson. Med. 2018, 80, 2188–2201. [Google Scholar] [CrossRef]
  118. Schlemper, J.; Caballero, J.; Hajnal, J.V.; Price, A.N.; Rueckert, D. A deep cascade of convolutional neural networks for dynamic MR image reconstruction. IEEE Trans. Med. Imaging 2017, 37, 491–503. [Google Scholar] [CrossRef] [Green Version]
  119. Wu, H.; Wu, Y.; Sun, L.; Cai, C.; Huang, Y.; Ding, X. A Deep Ensemble Network for Compressed Sensing MRI. In Proceedings of the International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 162–171. [Google Scholar]
  120. Liu, J.; Kuang, T.; Zhang, X. Image reconstruction by splitting deep learning regularization from iterative inversion. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 224–231. [Google Scholar]
  121. Wang, H.; Cheng, J.; Jia, S.; Qiu, Z.; Shi, C.; Zou, L.; Su, S.; Chang, Y.; Zhu, Y.; Ying, L.; et al. Accelerating MR imaging via deep Chambolle–Pock network. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 6818–6821. [Google Scholar]
  122. Sun, L.; Fan, Z.; Huang, Y.; Ding, X.; Paisley, J. Compressed sensing MRI using a recursive dilated network. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  123. Zeng, K.; Yang, Y.; Xiao, G.; Chen, Z. A Very Deep Densely Connected Network for Compressed Sensing MRI. IEEE Access 2019, 7, 85430–85439. [Google Scholar] [CrossRef]
  124. Ke, Z.; Wang, S.; Cheng, H.; Ying, L.; Liu, Q.; Zheng, H.; Liang, D. CRDN: Cascaded Residual Dense Networks for Dynamic MR Imaging with Edge-enhanced Loss Constraint. arXiv 2019, arXiv:1901.06111. [Google Scholar]
  125. Malkiel, I.; Ahn, S.; Taviani, V.; Menini, A.; Wolf, L.; Hardy, C.J. Conditional WGANs with Adaptive Gradient Balancing for Sparse MRI Reconstruction. arXiv 2019, arXiv:1905.00985. [Google Scholar]
  126. Huang, Q.; Yang, D.; Wu, P.; Qu, H.; Yi, J.; Metaxas, D. MRI reconstruction via cascaded channel-wise attention network. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1622–1626. [Google Scholar]
  127. Wang, P.; Chen, E.Z.; Chen, T.; Patel, V.M.; Sun, S. Pyramid Convolutional RNN for MRI Reconstruction. arXiv 2019, arXiv:1912.00543. [Google Scholar]
  128. Sun, L.; Fan, Z.; Fu, X.; Huang, Y.; Ding, X.; Paisley, J. A deep information sharing network for multi-contrast compressed sensing MRI reconstruction. IEEE Trans. Image Process. 2019, 28, 6141–6153. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  129. Jin, K.H.; Unser, M. 3D BBPConvNet to reconstruct parallel MRI. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, WA, USA, 4–7 April 2018; pp. 361–364. [Google Scholar]
  130. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; PMLR: Cambridge, MA, USA, 2017; pp. 214–223. [Google Scholar]
  131. Jiang, M.; Yuan, Z.; Yang, X.; Zhang, J.; Gong, Y.; Xia, L.; Li, T. Accelerating CS-MRI Reconstruction with Fine-Tuning Wasserstein Generative Adversarial Network. IEEE Access 2019, 7, 152347–152357. [Google Scholar] [CrossRef]
  132. Jun, Y.; Eo, T.; Shin, H.; Kim, T.; Lee, H.J.; Hwang, D. Parallel imaging in time-of-flight magnetic resonance angiography using deep multistream convolutional neural networks. Magn. Reson. Med. 2019, 81, 3840–3853. [Google Scholar] [CrossRef] [PubMed]
  133. Sandino, C.M.; Lai, P.; Vasanawala, S.S.; Cheng, J.Y. Accelerating cardiac cine MRI beyond compressed sensing using DL-ESPIRiT. arXiv 2019, arXiv:1911.05845. [Google Scholar]
  134. Pour Yazdanpanah, A.; Afacan, O.; Warfield, S. Deep Plug-and-Play Prior for Parallel MRI Reconstruction. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  135. Schlemper, J.; Duan, J.; Ouyang, C.; Qin, C.; Caballero, J.; Hajnal, J.V.; Rueckert, D. Data consistency networks for (calibration-less) accelerated parallel MR image reconstruction. arXiv 2019, arXiv:1909.11795. [Google Scholar]
  136. Zhou, Z.; Han, F.; Ghodrati, V.; Gao, Y.; Yin, W.; Yang, Y.; Hu, P. Parallel imaging and convolutional neural network combined fast MR image reconstruction: Applications in low-latency accelerated real-time imaging. Med. Phys. 2019, 46, 3399–3413. [Google Scholar] [CrossRef]
  137. Liu, Q.; Yang, Q.; Cheng, H.; Wang, S.; Zhang, M.; Liang, D. Highly undersampled magnetic resonance imaging reconstruction using autoencoding priors. Magn. Reson. Med. 2020, 83, 322–336. [Google Scholar] [CrossRef]
  138. Duan, J.; Schlemper, J.; Qin, C.; Ouyang, C.; Bai, W.; Biffi, C.; Bello, G.; Statton, B.; O’Regan, D.P.; Rueckert, D. VS-Net: Variable splitting network for accelerated parallel MRI reconstruction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 713–722. [Google Scholar]
  139. Wang, S.; Cheng, H.; Ying, L.; Xiao, T.; Ke, Z.; Liu, X.; Zheng, H.; Liang, D. DeepcomplexMRI: Exploiting deep residual network for fast parallel MR imaging with complex convolution. arXiv 2019, arXiv:1906.04359. [Google Scholar] [CrossRef] [Green Version]
  140. Meng, N.; Yang, Y.; Xu, Z.; Sun, J. A Prior Learning Network for Joint Image and Sensitivity Estimation in Parallel MR Imaging. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 732–740. [Google Scholar]
  141. Liu, R.; Zhang, Y.; Cheng, S.; Fan, X.; Luo, Z. A theoretically guaranteed deep optimization framework for robust compressive sensing mri. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4368–4375. [Google Scholar]
  142. Zhu, B.; Liu, J.Z.; Cauley, S.F.; Rosen, B.R.; Rosen, M.S. Image reconstruction by domain-transform manifold learning. Nature 2018, 555, 487–492. [Google Scholar] [CrossRef] [Green Version]
  143. Lee, D.; Yoo, J.; Ye, J.C. Deep residual learning for compressed sensing MRI. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 15–18. [Google Scholar]
  144. Lee, D.; Yoo, J.; Tak, S.; Ye, J.C. Deep residual learning for accelerated MRI using magnitude and phase networks. IEEE Trans. Biomed. Eng. 2018, 65, 1985–1995. [Google Scholar] [CrossRef] [Green Version]
  145. Gong, E.; Pauly, J.M.; Wintermark, M.; Zaharchuk, G. Deep learning enables reduced gadolinium dose for contrast-enhanced brain MRI. J. Magn. Reson. Imaging 2018, 48, 330–340. [Google Scholar] [CrossRef] [PubMed]
  146. Han, Y.; Yoo, J.; Kim, H.H.; Shin, H.J.; Sung, K.; Ye, J.C. Deep learning with domain adaptation for accelerated projection-reconstruction MR. Magn. Reson. Med. 2018, 80, 1189–1205. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  147. Ding, P.L.K.; Li, Z.; Zhou, Y.; Li, B. Deep residual dense U-Net for resolution enhancement in accelerated MRI acquisition. In Medical Imaging 2019: Image Processing; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10949, p. 109490F. [Google Scholar]
  148. Muckley, M.J.; Ades-Aron, B.; Papaioannou, A.; Lemberskiy, G.; Solomon, E.; Lui, Y.W.; Sodickson, D.K.; Fieremans, E.; Novikov, D.S.; Knoll, F. Training a Neural Network for Gibbs and Noise Removal in Diffusion MRI. arXiv 2019, arXiv:1905.04176. [Google Scholar] [CrossRef]
  149. Oksuz, I.; Clough, J.; Bustin, A.; Cruz, G.; Prieto, C.; Botnar, R.; Rueckert, D.; Schnabel, J.A.; King, A.P. Cardiac mr motion artefact correction from k-space using deep learning-based reconstruction. In International Workshop on Machine Learning for Medical Image Reconstruction; Springer: Berlin/Heidelberg, Germany, 2018; pp. 21–29. [Google Scholar]
  150. Seitzer, M.; Yang, G.; Schlemper, J.; Oktay, O.; Würfl, T.; Christlein, V.; Wong, T.; Mohiaddin, R.; Firmin, D.; Keegan, J.; et al. Adversarial and perceptual refinement for compressed sensing MRI reconstruction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 232–240. [Google Scholar]
  151. Dai, Y.; Zhuang, P. Compressed sensing MRI via a multi-scale dilated residual convolution network. Magn. Reson. Imaging 2019, 63, 93–104. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  152. Xiang, L.; Chen, Y.; Chang, W.; Zhan, Y.; Lin, W.; Wang, Q.; Shen, D. Deep-Learning-Based Multi-Modal Fusion for Fast MR Reconstruction. IEEE Trans. Biomed. Eng. 2018, 66, 2105–2114. [Google Scholar] [CrossRef] [PubMed]
  153. Zhou, W.; Du, H.; Mei, W.; Fang, L. Efficient Structurally-Strengthened Generative Adversarial Network for MRI Reconstruction. arXiv 2019, arXiv:1908.03858. [Google Scholar] [CrossRef]
  154. Han, Y.; Sunwoo, L.; Ye, J.C. k-space deep learning for accelerated MRI. IEEE Trans. Med. Imaging 2019, 39, 377–386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  155. Sun, L.; Fan, Z.; Huang, Y.; Ding, X.; Paisley, J. A Deep Error Correction Network for Compressed Sensing MRI. arXiv 2018, arXiv:1803.08763. [Google Scholar] [CrossRef] [Green Version]
  156. Eo, T.; Shin, H.; Kim, T.; Jun, Y.; Hwang, D. Translation of 1d inverse fourier transform of k-space to an image based on deep learning for accelerating magnetic resonance imaging. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 241–249. [Google Scholar]
  157. An, H.; Zhang, Y.J. A Structural Oriented Training Method for GAN Based Fast Compressed Sensing MRI. In Proceedings of the International Conference on Image and Graphics, Beijing, China, 23–25 August 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 483–494. [Google Scholar]
  158. Schlemper, J.; Yang, G.; Ferreira, P.; Scott, A.; McGill, L.A.; Khalique, Z.; Gorodezky, M.; Roehl, M.; Keegan, J.; Pennell, D.; et al. Stochastic deep compressive sensing for the reconstruction of diffusion tensor cardiac mri. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 295–303. [Google Scholar]
  159. Dar, S.U.H.; Özbey, M.; Çatlı, A.B.; Çukur, T. A transfer-learning approach for accelerated MRI using deep neural networks. arXiv 2017, arXiv:1710.02615. [Google Scholar] [CrossRef] [Green Version]
  160. Mardani, M.; Sun, Q.; Donoho, D.; Papyan, V.; Monajemi, H.; Vasanawala, S.; Pauly, J. Neural proximal gradient descent for compressive imaging. Adv. Neural Inf. Process. Syst. 2018, 31, 9573–9583. [Google Scholar]
  161. Liu, R.; Zhang, Y.; Cheng, S.; Luo, Z.; Fan, X. Converged Deep Framework Assembling Principled Modules for CS-MRI. arXiv 2019, arXiv:1910.13046. [Google Scholar]
  162. Chen, F.; Cheng, J.Y.; Taviani, V.; Sheth, V.R.; Brunsing, R.L.; Pauly, J.M.; Vasanawala, S.S. Data-driven self-calibration and reconstruction for non-cartesian wave-encoded single-shot fast spin echo using deep learning. J. Magn. Reson. Imaging 2020, 51, 841–853. [Google Scholar] [CrossRef] [PubMed]
  163. Biswas, S.; Aggarwal, H.K.; Jacob, M. Dynamic MRI using model-based deep learning and SToRM priors: MoDL-SToRM. Magn. Reson. Med. 2019, 82, 485–494. [Google Scholar] [CrossRef] [PubMed]
  164. Cheng, J.; Wang, H.; Ying, L.; Liang, D. Model learning: Primal dual networks for fast MR imaging. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 21–29. [Google Scholar]
  165. Cheng, J.; Wang, H.; Zhu, Y.; Liu, Q.; Ying, L.; Liang, D. Model-based Deep MR Imaging: The roadmap of generalizing compressed sensing model using deep learning. arXiv 2019, arXiv:1906.08143. [Google Scholar]
  166. Golbabaee, M.; Pirkl, C.M.; Menzel, M.I.; Buonincontri, G.; Gómez, P.A. Deep MR Fingerprinting with total-variation and low-rank subspace priors. arXiv 2019, arXiv:1902.10205. [Google Scholar]
  167. Zeng, W.; Peng, J.; Wang, S.; Liu, Q. A comparative study of CNN-based super-resolution methods in MRI reconstruction and its beyond. Signal Process. Image Commun. 2020, 81, 115701. [Google Scholar] [CrossRef]
  168. Zeng, D.Y.; Shaikh, J.; Nishimura, D.G.; Vasanawala, S.S.; Cheng, J.Y. Deep Residual Network for Off-Resonance Artifact Correction with Application to Pediatric Body Magnetic Resonance Angiography with 3D Cones. arXiv 2018, arXiv:1810.00072. [Google Scholar]
  169. Chun, Y.; Fessler, J.A. Deep BCD-net using identical encoding-decoding CNN structures for iterative image recovery. In Proceedings of the 2018 IEEE 13th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP), Aristi Village, Greece, 10–12 June 2018; pp. 1–5. [Google Scholar]
  170. Haber, E.; Lensink, K.; Triester, E.; Ruthotto, L. IMEXnet: A Forward Stable Deep Neural Network. arXiv 2019, arXiv:1903.02639. [Google Scholar]
  171. Narnhofer, D.; Hammernik, K.; Knoll, F.; Pock, T. Inverse GANs for accelerated MRI reconstruction. In Wavelets and Sparsity XVIII; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11138, p. 111381A. [Google Scholar]
  172. Tezcan, K.C.; Baumgartner, C.F.; Luechinger, R.; Pruessmann, K.P.; Konukoglu, E. MR image reconstruction using deep density priors. IEEE Trans. Med. Imaging 2018, 38, 1633–1642. [Google Scholar] [CrossRef]
  173. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  174. Salimans, T.; Karpathy, A.; Chen, X.; Kingma, D.P. Pixelcnn++: Improving the pixelcnn with discretized logistic mixture likelihood and other modifications. arXiv 2017, arXiv:1701.05517. [Google Scholar]
  175. Luo, G.; Zhao, N.; Jiang, W.; Cao, P. MRI Reconstruction Using Deep Bayesian Inference. arXiv 2019, arXiv:1909.01127. [Google Scholar]
  176. Yazdanpanah, A.P.; Afacan, O.; Warfield, S.K. Non-learning based deep parallel MRI reconstruction (NLDpMRI). In Medical Imaging 2019: Image Processing; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10949, p. 1094904. [Google Scholar]
  177. Jin, K.H.; Gupta, H.; Yerly, J.; Stuber, M.; Unser, M. Time-Dependent Deep Image Prior for Dynamic MRI. arXiv 2019, arXiv:1910.01684. [Google Scholar]
  178. Senouf, O.; Vedula, S.; Weiss, T.; Bronstein, A.; Michailovich, O.; Zibulevsky, M. Self-supervised learning of inverse problem solvers in medical imaging. In Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data; Springer: Berlin/Heidelberg, Germany, 2019; pp. 111–119. [Google Scholar]
  179. Yaman, B.; Hosseini, S.A.H.; Moeller, S.; Ellermann, J.; Uǧurbil, K.; Akçakaya, M. Self-Supervised Physics-Based Deep Learning MRI Reconstruction Without Fully-Sampled Data. arXiv 2019, arXiv:1910.09116. [Google Scholar]
  180. Yang, B.; Ying, L.; Tang, J. Artificial neural network enhanced Bayesian PET image reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1297–1309. [Google Scholar] [CrossRef]
  181. Xu, J.; Gong, E.; Pauly, J.; Zaharchuk, G. 200x low-dose PET reconstruction using deep learning. arXiv 2017, arXiv:1712.04119. [Google Scholar]
  182. Gong, K.; Guan, J.; Liu, C.C.; Qi, J. PET image denoising using a deep neural network through fine tuning. IEEE Trans. Radiat. Plasma Med. Sci. 2018, 3, 153–161. [Google Scholar] [CrossRef]
  183. Gong, Y.; Teng, Y.; Shan, H.; Xiao, T.; Li, M.; Liang, G.; Wang, G.; Wang, S. Parameter Constrained Transfer Learning for Low Dose PET Image Denoising. arXiv 2019, arXiv:1910.06749. [Google Scholar]
  184. Wang, Y.; Yu, B.; Wang, L.; Zu, C.; Lalush, D.S.; Lin, W.; Wu, X.; Zhou, J.; Shen, D.; Zhou, L. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. NeuroImage 2018, 174, 550–562. [Google Scholar] [CrossRef]
  185. Chen, K.T.; Gong, E.; de Carvalho Macruz, F.B.; Xu, J.; Boumis, A.; Khalighi, M.; Poston, K.L.; Sha, S.J.; Greicius, M.D.; Mormino, E.; et al. Ultra–low-dose 18F-florbetaben amyloid PET imaging using deep learning with multi-contrast MRI inputs. Radiology 2019, 290, 649–656. [Google Scholar] [CrossRef] [PubMed]
  186. Xiang, L.; Qiao, Y.; Nie, D.; An, L.; Lin, W.; Wang, Q.; Shen, D. Deep auto-context convolutional neural networks for standard-dose PET image estimation from low-dose PET/MRI. Neurocomputing 2017, 267, 406–416. [Google Scholar] [CrossRef] [PubMed]
  187. Liu, Z.; Chen, H.; Liu, H. Deep Learning Based Framework for Direct Reconstruction of PET Images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 48–56. [Google Scholar]
  188. Häggström, I.; Schmidtlein, C.R.; Campanella, G.; Fuchs, T.J. DeepPET: A deep encoder–decoder network for directly solving the PET image reconstruction inverse problem. Med. Image Anal. 2019, 54, 253–262. [Google Scholar] [CrossRef] [PubMed]
  189. Guazzo, A. Deep Learning for PET Imaging: From Denoising to Learned Primal-Dual Reconstruction. 2020. Available online: http://tesi.cab.unipd.it/64113/1/alessandro_guazzo_tesi.pdf (accessed on 11 January 2022).
  190. Kim, K.; Wu, D.; Gong, K.; Dutta, J.; Kim, J.H.; Son, Y.D.; Kim, H.K.; El Fakhri, G.; Li, Q. Penalized PET reconstruction using deep learning prior and local linear fitting. IEEE Trans. Med. Imaging 2018, 37, 1478–1487. [Google Scholar] [CrossRef]
  191. Gong, K.; Wu, D.; Kim, K.; Yang, J.; Sun, T.; El Fakhri, G.; Seo, Y.; Li, Q. MAPEM-Net: An unrolled neural network for Fully 3D PET image reconstruction. In 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11072, p. 110720O. [Google Scholar]
  192. Xie, Z.; Baikejiang, R.; Gong, K.; Zhang, X.; Qi, J. Generative adversarial networks based regularized image reconstruction for PET. In 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11072, p. 110720P. [Google Scholar]
  193. Gong, K.; Guan, J.; Kim, K.; Zhang, X.; Yang, J.; Seo, Y.; El Fakhri, G.; Qi, J.; Li, Q. Iterative PET image reconstruction using convolutional neural network representation. IEEE Trans. Med. Imaging 2018, 38, 675–685. [Google Scholar] [CrossRef]
  194. Gong, K.; Kim, K.; Cui, J.; Guo, N.; Catana, C.; Qi, J.; Li, Q. Learning personalized representation for inverse problems in medical imaging using deep neural network. arXiv 2018, arXiv:1807.01759. [Google Scholar]
  195. Cui, J.; Gong, K.; Guo, N.; Wu, C.; Meng, X.; Kim, K.; Zheng, K.; Wu, Z.; Fu, L.; Xu, B.; et al. PET image denoising using unsupervised deep learning. Eur. J. Nucl. Med. Mol. Imaging 2019, 46, 2780–2789. [Google Scholar] [CrossRef]
  196. Yokota, T.; Kawai, K.; Sakata, M.; Kimura, Y.; Hontani, H. Dynamic PET Image Reconstruction Using Nonnegative Matrix Factorization Incorporated with Deep Image Prior. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3126–3135. [Google Scholar]
  197. Hashimoto, F.; Ohba, H.; Ote, K.; Teramoto, A.; Tsukada, H. Dynamic PET image denoising using deep convolutional neural networks without prior training datasets. IEEE Access 2019, 7, 96594–96603. [Google Scholar] [CrossRef]
  198. Gong, K.; Catana, C.; Qi, J.; Li, Q. Direct patlak reconstruction from dynamic PET using unsupervised deep learning. In 15th International Meeting on Fully Three-Dimensional Image Reconstruction in Radiology and Nuclear Medicine; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11072, p. 110720R. [Google Scholar]
  199. Cui, J.; Gong, K.; Guo, N.; Kim, K.; Liu, H.; Li, Q. CT-guided PET parametric image reconstruction using deep neural network without prior training data. In Medical Imaging 2019: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 10948, p. 109480Z. [Google Scholar]
  200. Antun, V.; Renna, F.; Poon, C.; Adcock, B.; Hansen, A.C. On instabilities of deep learning in image reconstruction-Does AI come at a cost? arXiv 2019, arXiv:1902.05300. [Google Scholar]
  201. Ahmed, S.S.; Messali, Z.; Ouahabi, A.; Trepout, S.; Messaoudi, C.; Marco, S. Nonparametric denoising methods based on contourlet transform with sharp frequency localization: Application to low exposure time electron microscopy images. Entropy 2015, 17, 3461–3478. [Google Scholar] [CrossRef] [Green Version]
  202. Ouahabi, A. A review of wavelet denoising in medical imaging. In Proceedings of the 2013 8th International Workshop on Systems, Signal Processing and Their Applications (WoSSPA), Algiers, Algeria, 12–15 May 2013; pp. 19–26. [Google Scholar]
  203. Kumar, M.; Aggarwal, J.; Rani, A.; Stephan, T.; Shankar, A.; Mirjalili, S. Secure video communication using firefly optimization and visual cryptography. Artif. Intell. Rev. 2021, 1–21. [Google Scholar] [CrossRef]
  204. Dhasarathan, C.; Kumar, M.; Srivastava, A.K.; Al-Turjman, F.; Shankar, A.; Kumar, M. A bio-inspired privacy-preserving framework for healthcare systems. J. Supercomput. 2021, 77, 11099–11134. [Google Scholar] [CrossRef]
  205. Khan, S.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Yong, H.S.; Armghan, A.; Alenezi, F. Human Action Recognition: A Paradigm of Best Deep Learning Features Selection and Serial Based Extended Fusion. Sensors 2021, 21, 7941. [Google Scholar] [CrossRef] [PubMed]
  206. Syed, H.H.; Khan, M.A.; Tariq, U.; Armghan, A.; Alenezi, F.; Khan, J.A.; Rho, S.; Kadry, S.; Rajinikanth, V. A Rapid Artificial Intelligence-Based Computer-Aided Diagnosis System for COVID-19 Classification from CT Images. Behav. Neurol. 2021, 2021, 2560388. [Google Scholar] [CrossRef] [PubMed]
  207. Khan, M.A.; Alqahtani, A.; Khan, A.; Alsubai, S.; Binbusayyis, A.; Ch, M.; Yong, H.S.; Cha, J. Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection. Appl. Sci. 2022, 12, 593. [Google Scholar] [CrossRef]
  208. Haneche, H.; Boudraa, B.; Ouahabi, A. A new way to enhance speech signal based on compressed sensing. Measurement 2020, 151, 107117. [Google Scholar] [CrossRef]
Figure 1. A simple algorithm to solve the variational model. Blue arrows stand for the first step of iteration and the green arrows for the second one.
Figure 1. A simple algorithm to solve the variational model. Blue arrows stand for the first step of iteration and the green arrows for the second one.
Electronics 11 00586 g001
Figure 2. A simple algorithm to solve the Bayesian model. Two directions are denoted by blue and green arrows. The dash lines of different colors are used to represent the area of different prior probability. The probability value is from low to high when the color is changed from yellow to red.
Figure 2. A simple algorithm to solve the Bayesian model. Two directions are denoted by blue and green arrows. The dash lines of different colors are used to represent the area of different prior probability. The probability value is from low to high when the color is changed from yellow to red.
Electronics 11 00586 g002
Figure 3. An illustration for latent variable search of generative models. The dash line and scatter points represent M i m a g e . The blue arrows stand for r and generative model limits the movement direction of x in M i m a g e
Figure 3. An illustration for latent variable search of generative models. The dash line and scatter points represent M i m a g e . The blue arrows stand for r and generative model limits the movement direction of x in M i m a g e
Electronics 11 00586 g003
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xie, Y.; Li, Q. A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications. Electronics 2022, 11, 586. https://doi.org/10.3390/electronics11040586

AMA Style

Xie Y, Li Q. A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications. Electronics. 2022; 11(4):586. https://doi.org/10.3390/electronics11040586

Chicago/Turabian Style

Xie, Yutong, and Quanzheng Li. 2022. "A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications" Electronics 11, no. 4: 586. https://doi.org/10.3390/electronics11040586

APA Style

Xie, Y., & Li, Q. (2022). A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications. Electronics, 11(4), 586. https://doi.org/10.3390/electronics11040586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop