Next Article in Journal
Intelligent Clustering and Dynamic Incremental Learning to Generate Multi-Codebook Fuzzy Neural Network for Multi-Modal Data Classification
Previous Article in Journal
Robust Finite-Time Control of Linear System with Non-Differentiable Time-Varying Delay
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fixing Acceleration and Image Resolution Issues of Nuclear Magnetic Resonance

by
Krzysztof Malczewski
Faculty of Computing and Telecommunications, Poznan University of Technology, 60-965 Poznan, Poland
Symmetry 2020, 12(4), 681; https://doi.org/10.3390/sym12040681
Submission received: 27 March 2020 / Revised: 14 April 2020 / Accepted: 14 April 2020 / Published: 24 April 2020

Abstract

:
Lately, Magnetic Resonance scans have struggled with their own inherent limitations, such as spatial resolution as well as long examination times. A novel, rapid compressively-sensed magnetic resonance high-resolution image resolution algorithm is presented in this research paper. This technique addresses these two key issues by employing a highly-sparse sampling scheme and super-resolution reconstruction (SRR) method. Due to highly challenging requirements for the accuracy of diagnostic images registration, the presented technique exploits image priors, deblurring, parallel imaging, and a deformable human body motion analysis. Clinical trials as well as a phantom-based study have been conducted. It has been proven that the proposed algorithm can enhance image spatial resolution and reduce motion artefacts and scan times.

1. Introduction

Brain nuclear imaging has, in the last few years, become one of the most critical techniques to diagnose brain abnormalities. High-definition imaging with enough details has had its impact on medical imaging. As a result of these applications, the requirement for high quality visualisation is rapidly expanding. Yet, because of the constraints of inherent limitations of magnetic resonance imaging (MRI) scanners or issues related to the bandwidth in the transmission process, obtaining the high-definition head MR scans that meets the requirements for applications has been proving to be difficult. Efforts made to resolve this predicament have led to the improvement of developing studies in digital signal processing, especially in the field of image resolution enhancement. This area of interest has been especially has been researched in the latest years. The algorithm shown in this research paper enriches Iterative Back Projection (IBP) [1] and its improved version [2] in a few directions. It utilises a deformable image registration.
The SRR (Super-Resolution Image Reconstruction) has proven its worth in the field of various medical modalities [3,4,5]. Super Resolution Image Reconstruction is an example of an improperly modelled problem because of too-few low-resolution images. Numerous regularisation algorithms have been given to develop the inversion of this underdetermined problem. Freeman et al. [6] were the first to suggest that machine training procedures could be used to enhance the resolution of the frames.
By using the Markov random field (MRF), the authors defined the connection between the High- Resolution frame and the Low-Resolution frame. The preliminary estimate of the High-Resolution frame was accomplished by using interpolating polynomials. The missing upper frequencies of the High-Resolution frame were resolved using training technique and a preliminary estimate. Afterwards, the High-Resolution frame was calculated. Sun [7] has improved this technology by enhancing image features. The SR methods utilising the neural nets with convolution were proposed in the references [8,9]. This technique was employed in order to conduct single-multi-contrast SRR at the same time. It should be noted that the above-mentioned strategies rely on directories, which consist of large numbers of High-Resolution and Low-Resolution patch pairs. These large patch pair numbers amplify the complexity of computing. Moreover, fake high-frequency components exist, which are calculated by training procedures using external dictionaries. Despite all the addressed problems in the field of Super-Resolution, there are still shortcomings, and, in a certain sense, these problems are rather serious. Tao and Sajjadi [10,11] have gone further in improving the regular video super-resolution flowchart by using the previously estimated HR image to reconstruct the succeeding frame. The accuracy of these algorithms depends on the estimation of the motion vectors, which is time-consuming. Jo [12] has presented a curious slant on that matter. The point of his algorithm is to exploit an end-to-end deep neural nets to create dynamic upsampling filters and to derive a residual image. These two steps are calculated with respect to the local spatio-temporal knowledge obtained from neighbouring voxels. These procedures avoid using explicit motion compensation procedures. It should be noted, that deriving the temporal parameters exploiting 3D-kernels takes more time than in the case of 2D ones. Lately, Dong [13] suggested nesting deep convolutional neural networks at the core of SRR framework. Therefore, their report had also reflected the achievements made so far and highlighted the still-existing deficits, as a legacy from its predecessor to the one presented by them. Kim went deeper into the problem and applied [8,9,14] the residual network in order to improve network training procedures. Furthermore, Kim [15], Zhang [16,17] suggested using a residual scaling to create a network. Inspiring others to use the latest achievements in technology, the Deep Back-Projection Nets for SR implements down/up sampling structural layers to iteratively calculate an error feedback mechanism for error projection for all the stages of the algorithm.
A long examination phase is one of the key drawbacks of Magnetic Resonance Imaging. Despite this reason, making MRI data acquiring faster has been in the crosshairs of numerous scientists.
One of the possible scenarios of what could be developed is the change of phase encoding intervals in k-space filling. Unfortunately, weakened image quality can be a result. Yet, weakened image quality could be surmounted by utilising the proposed k-space sampling pattern. The sampling scheme suggested in this paper combines PROPELLER (Periodically Rotated Overlapping ParallEL Lines with Enhanced Reconstruction) sampling scheme [2] with Compressed Sensing and Partial Fourier, see Figure 1. The Partial Fourier framework claims that the phase changes gradually within the scanned object. Despite of the circumstances, numerous parallel imaging approaches such as SMASH (simultaneous acquisition of spatial harmonics) or SENSE (sensitivity encoding technique) [18], exploiting receiver-coil sensitivity prior knowledge is also useful in shortening the acquisition time [19]. The absent part of k-space could be salvaged by utilising its conjugate symmetry [20,21,22].
Consecutively, Compressed Sensing (CS) techniques [23], recently presented, may deliver further optimisation of sampling patterns. MRI revealed its relevance as the right place for applying CS frameworks. These methods can potentially accelerate the process to a higher level than any PI (Parallel Imaging) methods [23,24]. Numerous compressively-sensed methods have been brought into the limelight and tested. Considering the differences between several k-space sampling schemes, numerous efforts have been made to mutually utilise and associate unique, redundant knowledge achieved from various k-space sampling schemes. The combining of parallel imaging procedures with the Compressed Sensing methodology turns out to be more efficient in obtaining decreased image noise than with both the technologies used separately. In this work, the author presents a new MRI-associated method, which blends SR [2,3], deformable motion estimation with robust sampling trajectory pattern. The experimental results are promising and reveal the method’s true value.

2. MRI Sparse Sampling Schemes

The compressed sensing (CS) methodology was proposed to break the limitations of conventional sampling procedures [25,26]. The CS procedure exploits random undersampling and a sparsity-based template model image [27,28]. Over the last decade, many efforts have been made to improve diagnostic image reconstruction. Magnetic Resonance Imaging is a field of high priority because of its high examination charges. And, consequently, significant reduction of scanning times was a major reason for transplanting the CS to the field of Magnetic Resonance Imaging. Several researches have argued that the procedures such as k-t SPARSE, CS dynamic imaging [22] and k-t FOCUSS [29,30] images are compressible under a well-chosen sparsyfying transform. One of most important technological breakthroughs in medicine and healthcare was parallel imaging [22,31,32]. Currently, the majority of clinical scanners are loaded with parallel imaging methodology, and this way is the standard arrangement for numerous scanning schemes. It has been proven that scan time can be lowered by sampling a reduced number of phase encoding lines in frequency domain. Most modern clinical scanners collect input samples from several seperate receiver coil arrays [33].
Parallel imaging procedures utilise properties of these coil arrays to isolate aliased pixels in the image domain or to estimate missing k-space data exploiting knowledge of neighbouring k-space locations. Several approaches to the parallel imaging methods have been proposed [26,34]. The most prominent examples include SENSE, GRAPPA, and SPIRiT. They were all successfully applied in clinical practice. These procedures can be modified to consider irregular sampling schemes. Non-Cartesian sampling schemes deliver several valuable properties, that is, the appearance of incoherent aliasing artefacts. The latest improvements in concurrent multi-slice imaging are proposed, which utilise parallel imaging to separate images of numerous slices that have been obtained at the same time. Parallel imaging can also be applied to quicken three dimensional MRI, in which an adjacent volume is scanned rather than sequential slices. Another category of phase-constrained parallel imaging procedure makes use of both image magnitude and phase to improve reconstruction performance. Nevertheless, the robustness of compressively sensed MRI scanning procedures is still technically challenging. Numerous authors have noted that the pulse sequence is supposed to be wisely implemented to overcome any tendency of inconsistency, such as image resolution, number of frames/slices, contrast-to-noise ratio (CNR), and field of view. Meanwhile, other authors have proposed parallel imaging methods for handling dynamic Magnetic Resonance. In this way the techniques and methods which show the most effectiveness with certain types of procedures such as TSENSE and TGRAPPA [19] have been presented. They have proven their usefulness to capture 3–4 slices per heartbeat with relevant temporal and spatial resolution for diagnostic use, utilising commercially obtainable radiofrequency (RF) coil arrays. To enhance image resolution as well as the field of view, more complex and intricate methods are needed to achieve higher acceleration rates. Examples of this include procedures such as SENSE, GRAPPA and k-t-blast [35,36,37] to ensure the precision of exploiting spatiotemporal correlations in the dynamic Magnetic Resonance Imaging data either alone or in combination with coil sensitivity information. Such data sampling, as expect, gives rise, to reducing data overlapping, which delivers high acceleration rates. The most significant disadvantage is the need of dynamic learning data to set up an aliasing pattern in the frequency domain, which minimises the resulting acceleration rate. In opposite to these techniques, the CS framework relies on the assumption that a sparsely sampled frame in a known transform domain can be calculated using randomly undersampled k-space data [38] and does not need any training data.
Figure 1. Combining Compressed Sensing (CS) with exploiting the Hermitian symmetry property produces effective sampling pattern. On the right: the two k-space locations marked in yellow colour on the graph, mirror images across the origin of k-space, have identical amplitudes but opposite phases.
Figure 1. Combining Compressed Sensing (CS) with exploiting the Hermitian symmetry property produces effective sampling pattern. On the right: the two k-space locations marked in yellow colour on the graph, mirror images across the origin of k-space, have identical amplitudes but opposite phases.
Symmetry 12 00681 g001

3. The Proposed Algorithm

Most SRR algorithms incline to produce similar results because of oversimplified motion trajectories which do not display real biological behaviour. This is crucial in all instances of patient motion within an image, such as respiratory, heart motion, circulation, and other potentially misregistration-causing artefacts. A deformable, groupwise non-rigid image registration method for motion compensation is utilised in the algorithm being presented in this paper [39,40].
Medical image processing has experienced a variety of difficulties in the past including distortion removal, spatial resolution and examination times [41,42]. Mathematically speaking, all these aspects lead to forming a cost function, which is crucial to get accurate and better predictions.This registration algorithm had proven its competency in the Computed Tomography [43], PET/MRI as well as DW-MRI fields. The proposed SRR algorithm concurrently uses High-Resolution sparsity priors, deformable motion registration parameters as well as the MAP (maximum a posteriori) for blur estimation. The proposed SRR algorithm minimises [39,40] the cost function consisting of the HR input image Γ H , motion fields { w i } , noise level { θ i } as well blur kernel B. The algorithm gets started with using an initial guess which yields convergence to a solution, see the pseudocode below and Figure 2 and Figure 3.
Figure 2. An initial guess estimation yields convergence to a solution.
Figure 2. An initial guess estimation yields convergence to a solution.
Symmetry 12 00681 g002
Input: set of Low-resolution input Magnetic Resonance Scans /each is reconstructed from a compressively sensed k-space’s blade.
  • high-resolution estimate produced by → the submethod, see Figure 2.
  • iterate until convergence | Γ L n Γ L n , observed | < ε
    • Estimate noise θ i = α + N q 1 β + N q x ¯
    • Calculated deformable image registration parameters and utilise them to align an image grid
    • Estimate blur kernel B x = a r g min θ 0 B x A M y B x Γ L n + ξ B x
    • Enhance the high-resolution estimate Γ H n
    • Repeat steps a-d until convergence, see Figure 3
Output: high-resolution Magnetic Scan.
Figure 3. The proposed algorithm flowchart.
Figure 3. The proposed algorithm flowchart.
Symmetry 12 00681 g003
The procedure performs image upscaling regarding the motion fields to produce a current HR image estimate. The proposed algorithm utilises a smooth kernel, which blurs the resulting image. The image is subsampled and noised to form the set of observed low-resolution images. This iterative technique is continued until the solution converges towards values that change by less than some specified tolerance threshold between successive iterations [7]. The proposed SRR algorithm can be expressed in the following way:
Γ ^ H n = a r g min θ 0 Γ H n D B Γ H n Γ L n , 0 + g Γ H n + i = N , i 0 N D B w i Γ H n Γ L n , 0 .
Note that the Γ L n , 0 represents the estimated reference initial guess related to n-th set of “compressed” PROPELLER blades, Γ L n , j is the i-th obtained LR frame with regard to degradation parameters, Γ L n is HR estimate, is the gradient operator, Γ ^ H n is the n-th HR image estimate D , B , w i denote the down-sampling, blur kernel, deformable image registration vectors, individually. To be able to employ gradient optimisation we need to replace L1 norms with their differentiable approximations.
It should be noted that the deformable motion estimation is a prominent example of a highly non- convex optimisation problem, which is associated with several conditions that must be satisfied [44]. Moreover, this process is highly ill-conditioned tending to sensitivity regarding measurements and model errors. The motion estimation algorithm adopted in this paper is globally optimal deformable registration. That algorithm avoids using continuous optimisation because it may be prone to local minima. To overcome this advantage, discrete optimisation was applied. The frame grid is modelled as a minimum spanning tree. Instead of using gradients to find a global optimum of the cost function, it can be found rapidly using dynamic programming, which enforces the smoothness of the deformations.
The Markov Random Field (MRF) labelling is used to perform discrete optimisation, see Figure 4. In the algorithm adopted in this paper a spanning tree with minimum total edge costs is derived. The nodes i P refer to pixels (or group of pixels) and for each node, a collection of hidden, and corresponding to the motion fields, labels are expressed as w i l = { f i l , g i l , h i l } . The algorithm, more specifically, the energy function to be optimised consists of two terms: that is, the data cost S and the pair-wise regularisation cost R ( w i l , w i m ) for all the nodes l s associated with the nodes m s :
E ( w i ) = i P S ( w i l ) + κ l , m N R ( w i l , w i m ) .
The cost function shown above estimates the inter-pixel correlation of a pair of images being compared. Specifically, this term does not depend on the displacements of its neighbours.
The κ refers to a weighting parameter, which controls the influence of the regularisation term. The first component in Equation (2) denotes the data term; the second one is the regularisation term.
Figure 4. The Magnetic Resonance image registration. This procedure is applied to the entire set of the Low-Resolution images.
Figure 4. The Magnetic Resonance image registration. This procedure is applied to the entire set of the Low-Resolution images.
Symmetry 12 00681 g004

4. Results

In the experimental studies, all the raw data signals are appropriately compressed with the sampling rates: 25%, 40% and 60% of the fully sampled k-spaces.
In order to measure the performance of the proposed algorithm, both laboratory phantom studies and an in-vivo assessment were performed. The purpose of the experiment was to compare the effectiveness of various ways of obtaining a compressively-sensed input. In a study of the effects of various compressed-sensing ratio on test performance, the test performance of the proposed Super-Resolution Image Reconstruction method was tested. Moreover, several MRI k-space sampling patterns have been compared. Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 show the achieved results. It must be emphasised that combining Compressed Sensing with Hermitian symmetry property, as well as Partial Fourier allows the shortening of k-space filling when compared to the different k-space sampling schemes, see Figure 1.
Figure 5. The phantom images based on experiment results. See from the upper row: the unprocessed image corrupted by simulated shift and the reconstructed images with various sampling rates (varying from 25% to 40% and 60% of samples of the ground truth).
Figure 5. The phantom images based on experiment results. See from the upper row: the unprocessed image corrupted by simulated shift and the reconstructed images with various sampling rates (varying from 25% to 40% and 60% of samples of the ground truth).
Symmetry 12 00681 g005
Figure 6. A brain imaging example. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 1 for the PSNR values at other compression ratios.
Figure 6. A brain imaging example. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 1 for the PSNR values at other compression ratios.
Symmetry 12 00681 g006
Table 1. Stats of the Peak signal-to-noise ratio (PSNR) metrics for Figure 7. MAE abbreviation stands for Mean Average Error.
Table 1. Stats of the Peak signal-to-noise ratio (PSNR) metrics for Figure 7. MAE abbreviation stands for Mean Average Error.
Reconstruction MethodNMMAESDtp
down-sampled without image registration10021.1620.040.011.0940.276
down-sampled with image registration10024.5219.340.01−0.7790.438
motion corrected regular sampling scheme
(without subsampling applied)
10022.3618.040.010.1850.854
B-spline cubic interpolation /image registration applied/10023.3118.010.010.1840.274
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor10026.3317.220.01−0.3210.432
Enhanced deep residual networks for single image super-resolution10029.1416.550.01−0.3620,412
Image super-resolution using very deep residual channel
attention networks
10028.7515.510.01−0.4160.437
Residual dense network for image SR10029.8814.630.01−0.5410.554
the proposed method10030.3914.020.01−0.5880.558
Figure 7. The brain imaging results. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 2 for the PSNR values at other compression ratios.
Figure 7. The brain imaging results. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 2 for the PSNR values at other compression ratios.
Symmetry 12 00681 g007
Table 2. Stats of the Image Enhancement Metric (IEM) metrics for Figure 7.
Table 2. Stats of the Image Enhancement Metric (IEM) metrics for Figure 7.
Reconstruction MethodNMSDtp
down-sampled without image registration1001.760.011.2220.225
down-sampled with image registration1001.990.010.5050.615
motion corrected regular sampling scheme
(without subsampling applied)
1002.670.011.8480.068
B-spline cubic interpolation /image registration applied/1002.010.010.1840.273
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor1002.330.01−0.3200.436
Enhanced deep residual networks for single image super-resolution1002.140.01−0.3610.411
Image super-resolution using very deep residual channel
attention networks
1002.450.01−0.4110.431
Residual dense network for image SR1002.880.01−0.5430.552
the proposed method1003.990.00−1.9010.061
Figure 8. The abdominal image processing. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 3 for the PSNR values at other compression ratios.
Figure 8. The abdominal image processing. From left to right: A: Image reconstructed from partially sampled PROPELLER blade, B:Cartesian sampling grid without image registration applied (with no downsampling applied), C: B-spline Cubic interpolation, D: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], E: Enhanced deep residual networks for single image super-resolution [14], F: Image super-resolution using very deep residual channel attention networks [16], G: Residual dense network for image super-resolution [15], H: super-resolution with proposed sampling scheme and motion compensation (the proposed algorithm). Compression ratio is 50%. Please see Table 3 for the PSNR values at other compression ratios.
Symmetry 12 00681 g008
Table 3. Stats of the PSNR metrics for Figure 17. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Table 3. Stats of the PSNR metrics for Figure 17. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Reconstruction MethodNMMAESDtp
Image reconstructed from partially sampled
PROPELLER blade
10019.1619.220.011.6540.101
no motion corrected regular sampling scheme
(with no downsampling applied)
10026.2118.340.010.6720.503
B-spline cubic interpolation /image registration applied/10023.3117.890.010.1830.273
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor10026,2217.020.01−0.3540.431
Enhanced deep residual networks for single image super-resolution10029.6516.440.01−0.3840,411
Image super-resolution using very deep residual channel
attention networks
10028.6616.010.01−0.4660.436
Residual dense network for image SR10029.0015.550.01−0.5650.554
the proposed method10029.2814.070.01−1.0020.554
Figure 9. The Shepp-Logan phantom results comparison. From the left: the PROPELLER sampling reconstruction output (PSNR = 29.84 dB, IEM = 2.04), the proposed algorithm result with enhanced resolution (PSNR = 34.58 dB, IEM = 3.64). The lower row shows detailed images. Please see Table 4 for the PSNR values at other compression ratios.
Figure 9. The Shepp-Logan phantom results comparison. From the left: the PROPELLER sampling reconstruction output (PSNR = 29.84 dB, IEM = 2.04), the proposed algorithm result with enhanced resolution (PSNR = 34.58 dB, IEM = 3.64). The lower row shows detailed images. Please see Table 4 for the PSNR values at other compression ratios.
Symmetry 12 00681 g009
Table 4. Stats of the PSNR metrics for Table 5 at different CS ratio. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Table 4. Stats of the PSNR metrics for Table 5 at different CS ratio. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
CS Quality [%]NMMAESDt(99)p
2010018.7620.010.010.6470.519
4010025.6218.080.010.7990.426
5010030.3915.010.011.8480.068
6010031.1614.040.011.2220.225
Table 5. Stats of the PSNR metrics for Figure 12. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Table 5. Stats of the PSNR metrics for Figure 12. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Reconstruction MethodNMMAESDtp
Image reconstructed from partially sampled
PROPELLER blade10018.2320.010.01−0.3000.765
no motion corrected regular sampling scheme
(with no downsampling applied)10026.8619.720.01−1.5540.121
B-spline cubic interpolation /image registration applied/10023.3018.910.010.1810.275
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor10026.311.230.01−0.3230.437
Enhanced deep residual networks for single image super-resolution10029.4116.000.01−0.3670.411
Image super-resolution using very deep residual channel
attention networks10028.1215.610.01−0.4120.432
Residual dense network for image SR10029.9315.010.01−0.5420.558
the proposed method10036.2214.320.011.3470.181
Figure 10. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 7.
Figure 10. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 7.
Symmetry 12 00681 g010
Figure 11. The beeswarm* plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 7. *The so-called bee swarm plot gives a better representation of the distribution of values, but it does not scale well to large numbers of observations.
Figure 11. The beeswarm* plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 7. *The so-called bee swarm plot gives a better representation of the distribution of values, but it does not scale well to large numbers of observations.
Symmetry 12 00681 g011
Figure 12. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 7 (Table 4 illustrated).
Figure 12. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 7 (Table 4 illustrated).
Symmetry 12 00681 g012
Figure 13. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 7 (Table 6 illustrated).
Figure 13. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 7 (Table 6 illustrated).
Symmetry 12 00681 g013
Table 6. Stats of the IEM metrics for Figure 7 at different CS ratio.
Table 6. Stats of the IEM metrics for Figure 7 at different CS ratio.
CS Quality [%]NIEMSDt(99)p
201001.790.00−1.4210.158
401001.880.011.6540.101
501003.990.01−1.1790.241
601004.310.000.0001.000
Figure 14. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 12.
Figure 14. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 12.
Symmetry 12 00681 g014
Figure 15. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 12.
Figure 15. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 12.
Symmetry 12 00681 g015
Figure 16. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 12.
Figure 16. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 12.
Symmetry 12 00681 g016
Figure 17. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 12.
Figure 17. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 12.
Symmetry 12 00681 g017
Figure 18. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 18.
Figure 18. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 18.
Symmetry 12 00681 g018
Figure 19. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 17.
Figure 19. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 17.
Symmetry 12 00681 g019
Figure 20. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 17.
Figure 20. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 17.
Symmetry 12 00681 g020
Figure 21. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 17.
Figure 21. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 17.
Symmetry 12 00681 g021
It must be emphasised that combining Compressed Sensing with Hermitian symmetry property, as well as Partial Fourier allows the shortening of k-space filling when compared to the different k-space sampling schemes. This paper focuses on fusing the super-resolution image reconstruction with k-space sparse sampling of MRI scanners. Moreover, the phantom studies were concentrated on the inverse problems of compressed sensing for MRI, see Figure 9. In this part of the experiment the Shepp-Logan phantom data Figure 6 and Figure 22 were reconstructed from sets of sparse projections. To model the sparsity, PROPELLER blades have been sensitively compressed with 30 and twelve radial lines in frequency domain as well as reconstruction from reduced-angle projections, with a contracted subset of 50 projections within a 90-degree aperture. The unmodified PROPELLER sampling scheme has been compared with the proposed one, see Figure 23. Furthermore, the ground truth MR images meshes were warped using nonrigid, deformable transformations. Lastly, Gaussian blur kernel and noise and downsampling were applied to such processed images to simulate real MR images.
Figure 22. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 22.
Figure 22. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the PSNR scores for all the methods on each dataset from Figure 22.
Symmetry 12 00681 g022
Figure 23. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 22.
Figure 23. The bee swarm plots are one-dimensional scatter plots showing the “stripcharts” of the IEM scores for all the methods on each dataset from Figure 22.
Symmetry 12 00681 g023

5. Discussion

For over nearly two decades, SR procedures have effectively been exploited to enhance the spatial resolution of diagnostic images after k-spaces data are collected, thus making the doctor’s diagnosis easier. The variety of applications and methods has grown ever since, especially in the MRI modality, exposing the interest of the community to such post- processing. MRI, CT, PET and hybrid techniques are still suffering from insufficient spatial resolution, contrast issues, visual noise scattering and blurring produced by motion artefacts. These underlying issues can lead to problems in identifying abnormalities. Reducing scanning time is a serious challenge for many medical imaging techniques. Compressed Sensing (CS) theory delivers an appealing framework to address this inconvenience since it provides theoretical guarantees on the reconstruction of sparse signals by projection on a low dimensional linear subspace. Numerous adjustments here are presented and proven to be efficient for enhancing MRI spatial resolution while reducing acquisition lengths. The proposed technique minimises artefacts produced by highly sparse data, even in the influence of misregistration artefacts. In order to expedite the process of k-space filling and enhance spatial resolution, the algorithm combines several sub-techniques, that is, compressive-sensing, Poisson Disc sampling and Partial Fourier with SR technique. This combination allowed for improving both image definition and time consumption. Furthermore, improved upper frequencies provides better edge delineation. The proposed algorithm can be directly implemented to MR scanners without any hardware modifications. Moreover, it has been proven that the implemented technique produces enhanced and sharper shapes. It really minimises the risk of misdiagnosis. The financial aspects in healthcare must be addressed and sufficient value to all participants are supposed to justify their use. Apart from enhancing the spatial resolution, this technique can be useful in addressing misregistration issues. Phantom-based studies as well in-vivo experiments have proven to be successful in reducing examination time. The achieved results show an improvement of definition and readability of the MR images, see Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21, Figure 22 and Figure 23. The techniques used to identify abnormalities that are claimed to be potentially malignant or pre-malignant expose a higher capability to spot them due to the improved image quality parameters. Furthermore, the achievements have been validated by PSNR and IEM metrics [45] that can measure medical image quality with great competence. In more extended studies numerous scanning techniques were investigated. The proposed algorithm has been compared with SENSE, GRAPPA as well as unmodified PROPELLER. Despite reasonable results achieved using competitive algorithms, the proposed algorithms offered the shortest scanning time, see Table 7. In order to get a quantitative assessment, the peak signal-to-noise ratio (PSNR) and the mean absolute error (MAE) have been utilised. Both the smaller MAE or higher PSNR confirmed robustness of the applied algorithm. All the most commonly applied image quality metrics, including the IEM, confirmed the weight of the results. The proposed algorithm was compared with 4 state-of-the-art SRRs: Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor [45], Residual dense network for image super-resolution [15], Enhanced deep residual networks for single image super-resolution [14] and Image super-resolution using very deep residual channel attention networks [16]. L1-cost regularisation is applied to learn the nets of the following procedures—Enhanced deep residual networks for single image super-resolution, Residual dense network for image super-resolution and Image super-resolution using very deep residual channel attention networks. The compressively-sensed, Super-Resolution images have achieved the highest IEM values [46]. it is obvious to see that the compression ratios highly affect the PSNR scores (see Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18). It is worth being underlined that satisfying results occurred for halved sampling spaces. The mean PSNR and IEM values are used to compare the results. Each simulation was run N = 100 times. The signed rank test was applied to all the image quality metrics to verify if the null hypothesis that the central tendency of the difference was zero at numerous acceleration rates. All the statistical tests were performed using The R Project for Statistical Computing, see Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21, Figure 22 and Figure 23. A separate group t-student‘s test was conducted in order to compare the PSNR mean scores between 2 seperate sets, with a paired t-test. The significance test, Student’s t-test was carried out on the PSNR and are found that the probability validates the method’s performance. The results are exposed in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17, Table 18 and Table 19. The achieved p-values have proven high statistical significance. Future research will be concentrated on testing competitive solutions which operate using Discrete Shearlet Transform because of its ability to be a good candidate for MRI sparse sampling. Presently, the algorithm is being validated via testing using DW-MRI/PET scanners.
Table 7. Comparison of the scanning parameters
Table 7. Comparison of the scanning parameters
Scanning PatternTRTEFOVVoxel (mm)Total Scan Duration (s)p
PROPELLER12001802900.96/0.96/1.003600.159
SENSE12001802900.96/0.96/1.003530.226
GRAPPA12001802900.96/0.96/1.003200.136
the proposed algorithm12001802900.96/0.96/1.001120.103
Table 8. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
Table 8. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
CS Ratio [%]PSNR [dB]MAEIEM
2018.7619.821.79
4025.6218.011.88
5030.3915.663.99
6031.1614.204.31
Table 9. Stats of the IEM metrics for Figure 12. M denotes mean of observed IEMs
Table 9. Stats of the IEM metrics for Figure 12. M denotes mean of observed IEMs
Reconstruction MethodNMSDtp
Image reconstructed from partially sampled
PROPELLER blade1001.560.011.1330.260
no motion corrected regular sampling scheme
(with no downsampling applied)1003.120.010.2940.070
B-spline cubic interpolation /image registration applied/1002.620.011.8420.061
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor1002.110.010.1830.275
Enhanced deep residual networks for single image super-resolution1002.320.01−0.3270.436
Image super-resolution using very deep residual channel
attention networks1002.150.01−0.3660.412
Residual dense network for image SR1002.430.01−0.4120.432
the proposed method1002.820.01−0.510.551
Image reconstructed from partially sampled
PROPELLER blade1003.890.01−0.3710.202
Table 10. The performance of the proposed algorithm at different CS ratio for Figure 12. MAE abbreviation stands for Mean Average Error.
Table 10. The performance of the proposed algorithm at different CS ratio for Figure 12. MAE abbreviation stands for Mean Average Error.
CS Ratio [%]PSNR [dB]MAEIEM
2018.4420.011.75
4028.4218.031.92
5036.2216.053.89
6038.1114.554.31
Table 11. Stats of the PSNR metrics for Figure 12. M denotes mean of observed PSNRs.
Table 11. Stats of the PSNR metrics for Figure 12. M denotes mean of observed PSNRs.
CS Quality [%]NMSDt(99)p
2010018.440.010.1390.889
4010028.420.010.7280.469
5010036.220.011.7890.077
6010038.110.011.8300.070
Table 12. Stats of the IEM metrics for Figure 12. M denotes mean of observed IEMs
Table 12. Stats of the IEM metrics for Figure 12. M denotes mean of observed IEMs
CS Quality [%]NMSDt(99)p
201001.750.011.0000.329
401001.920.01−0.8650.389
501003.890.01−0.6720.503
601004.310.010.9610.339
Table 13. Stats of the IEM metrics for Figure 17. M denotes mean of observed IEMs.
Table 13. Stats of the IEM metrics for Figure 17. M denotes mean of observed IEMs.
Reconstruction MethodNMSDtp
Image reconstructed from partially sampled
PROPELLER blade1001.560.011.1330.260
no motion corrected regular sampling scheme
(with no downsampling applied)1003.120.010.2940.770
B-spline cubic interpolation /image registration applied/1002.620.011.8220.021
Non-Rigid Multi-Modal 3D Medical Image Registration
Based on Foveated Modality Independent Neighbourhood Descriptor1002.100.010.1230.225
Enhanced deep residual networks for single image super-resolution1002.310.01−0.3170.426
Image super-resolution using very deep residual channel
attention networks1002.120.01−0.3460.442
Residual dense network for image SR1002.490.01−0.4620.412
the proposed method1002.800.01−0.5110.521
Image reconstructed from partially sampled
PROPELLER blade1003.890.01−0.3710.202
Table 14. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
Table 14. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
CS Ratio [%]PSNR [dB]MAEIEM
2019.3120.011.67
4024.1418.231.91
5029.2816.023.68
6031.0815.654.19
Table 15. Stats of the PSNR metrics for Figure 17. M denotes mean of observed PSNRs.
Table 15. Stats of the PSNR metrics for Figure 17. M denotes mean of observed PSNRs.
CS Quality [%]NMSDt(99)p
2010019.310.010.5420.589
4010024.140.01−0.7130.478
5010029.280.01−1.0440.299
6010031.080.01−1.0210.310
Table 16. Stats of the IEM metrics for Figure 17. M denotes mean of observed IEMs.
Table 16. Stats of the IEM metrics for Figure 17. M denotes mean of observed IEMs.
CS Quality [%]NMSDt(99)p
201001.670.01−0.5750.566
401001.910.01−0.1340.894
501003.680.010.5420.589
601004.190.010.5880.558
Table 17. Stats of the PSNR metrics for Figure 22. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Table 17. Stats of the PSNR metrics for Figure 22. M denotes mean of observed PSNRs. MAE abbreviation stands for Mean Average Error.
Reconstruction/Sampling AlgorithmNMMAESDt(99)p
the PROPELLER sampling reconstruction10029.8418.220.01−1.8810.063
the proposed algorithm10034.5815..010.001.1490.253
Table 18. Stats of the IEM metrics for Figure 22. M denotes mean of observed IEMs
Table 18. Stats of the IEM metrics for Figure 22. M denotes mean of observed IEMs
Reconstruction/Sampling AlgorithmNMSDt(99)p
the PROPELLER sampling reconstruction1002.040.010.1390.889
the proposed algorithm1003.640.00−1.2280.222
Table 19. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
Table 19. The performance of the proposed algorithm at different CS ratio. MAE abbreviation stands for Mean Average Error.
CS Ratio [%]PSNR [dB]MAEIEM
2021.0420.221.88
4027.4518.231.92
5034.5815.453.64
6036.0115.014.17

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest. In the past five years, I have not received reimbursements, fees, funding, or any salaries from any organisation that may in any way gain or lose financially from the publication of this manuscript. I do not hold any stocks or shares in an organisation that may in any way gain or lose financially from the publication of this manuscript. I do not hold and I am not currently applying for any patents relating to the content of the manuscript. I do not have any other financial competing interests.

References

  1. Irani, M.; Peleg, S. Improving resolution by image registration. CVGIP Graph. Models Image Process. 1991, 53, 231–239. [Google Scholar] [CrossRef]
  2. Malczewski, K.; Stasinski, R. High-resolution MRI image reconstruction from a PROPELLER data set of samples. Int. J. Funct. Inform. Pers. Med. 2008, 1, 311–320. [Google Scholar] [CrossRef]
  3. Malczewski, K. Breaking The Resolution Limit In Medical Imaging Modalities. In Proceedings of the 2012 International Conference on Image Processing, Computer Vision, Worldcomp and Pattern Recognition (IPCV’12), Las Vegas, NV, USA, 16–19 July 2012. [Google Scholar]
  4. Malczewski, K.; Stasinsk, R. Super Resolution for Multimedia, Image, and Video Processing Applications. In Studies in Computational Intelligence; Springer: Berlin, Germany, 2009; Volume 231. [Google Scholar]
  5. Kennedy, J.A.; Israel, O.; Frenkel, A.; Bar-Shalom, R.; Azhari, H. Super-resolution in PET imaging. IEEE Trans. Med. Imaging 2006, 25, 137–147. [Google Scholar] [CrossRef]
  6. Freeman, W.T.; Pasztor, E.C. Learning to Estimate Scenes from Images. In Advances in Neural Information Processing Systems; Kearns, M.S., Solla, S.A., Cohn, D.A., Eds.; MIT Press: Cambridge, MA, USA, 1999; Volume 11, pp. 775–781. [Google Scholar]
  7. Liu, C.; Sun, D. On Bayesian Adaptive Video Super Resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 346–360. [Google Scholar] [CrossRef] [Green Version]
  8. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  9. Kim, J.; Lee, J.K.; Lee, K.M. Deeply recursive convolutional network for image super-resolution. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  10. Tao, X.; Gao, H.; Liao, R.; Wang, J.; Jia, J. Detail-revealing deep video super-resolution. In Proceedings of the IEEE International Conference on Computer Vision (ICCV 2017), Venice, Italy, 22–29 October 2017; pp. 4482–4490. [Google Scholar]
  11. Sajjadi, M.S.M.; Vemulapalli, R.; Brown, M. Frame-recurrent video super-resolution. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2018), Salt Lake City, UT, USA, 18–22 June 2018; pp. 6626–6634. [Google Scholar]
  12. Younghyun, J.; Seoung, W.O.; Jaeyeon, K.; Seon, J.K. Deep video super-resolution network using dynamic upsampling filters without explicit motion compensation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  13. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  14. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced deep residual networks for single image super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  15. Zhang, Y.; Tian, Y.; Kong, Y.; Fu, B.Y. Residual dense network for image super-resolution. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar]
  16. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; Part VII. pp. 294–310. [Google Scholar]
  17. Haris, M.; Shakhnarovich, G.; Ukita, N. Deep back-projection networks for super-resolution. In Proceedings of the 2018 IEEE Conference on Computer Vision and PatternRecognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1664–1673. [Google Scholar]
  18. Pruessmann, K.P.; Weiger, M.; Scheidegger, M.B.; Boesiger, P. SENSE: Sensitivity encoding for fast MRI. Magn. Reson. Med. 1999, 42, 952–962. [Google Scholar] [CrossRef]
  19. Ding, Y.; Chung, Y.C.; Jekic, M.; Simonetti, O.P. A new approach to auto-calibrated dynamic parallel imaging based on the Karhunen-Loeve transform: KL-TSENSE and KL-TGRAPPA. Magn. Reson. Med. 2011, 65, 1786–1792. [Google Scholar] [CrossRef] [Green Version]
  20. Küstner, T.; Würslin, C.; Gatidis, S.; Martirosian, P.; Nikolaou, K.; Schwenzer, N.F.; Schick, F.; Yang, B.; Schmidt, H. MR image reconstruction using a combination of Compressed Sensing and partial Fourier acquisition: ESPReSSo. IEEE Trans. Med. Imaging 2016, 35, 2447–2458. [Google Scholar] [CrossRef]
  21. Ding, Y.; Chung, Y.C.; Jekic, M.; Simonetti, O.P. A method to assess spatially variant noise in dynamic MR image series. Magn. Reson. Med. 2010, 63, 782–789. [Google Scholar] [CrossRef] [PubMed]
  22. Kim, D.; Dyvorne, H.A.; Otazo, R.; Feng, L.; Sodickson, D.K.; Lee, V.S. Accelerated phase-contrast cine MRI using k-t SPARSE-SENSE. Magn. Reson. Med. 2012, 67, 1054–1064. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Pedersen, H.; Kozerke, S.; Ringgaard, S.; Nehrke, K.; Kim, W. k-t PCA: Temporally constrained k-t BLAST reconstruction using principal component analysis. Magn. Reson. Imaging 2009, 63, 706–716. [Google Scholar]
  24. Griswold, M.A.; Jakob, P.M.; Heidemann, R.M.; Nittka, M.; Jellus, V.; Wang, J.; Kiefer, B.; Haase, A. Generalized autocalibrating partially parallel acquisitions (GRAPPA). Magn. Reson. Med. 2002, 47, 1202–1210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Noll, D.C.; Nishimura, D.G.; Macovski, A. Homodyne detection in magnetic resonance imaging. IEEE Trans. Med. Imaging 1991, 10, 154–163. [Google Scholar] [CrossRef] [PubMed]
  26. McGibney, G.; Smith, M.R.; Nichols, S.T.; Crawley, A. Quantitative evaluation of several partial Fourier Reconstruction methods used in MRI. Magn. Reson. Med. 1993, 30, 51–59. [Google Scholar] [CrossRef] [PubMed]
  27. Davenport, M. The Fundamentals of Compressive Sensing. IEEE Signal Processing Society Online Tutorial Library, 12 April 2013. [Google Scholar]
  28. Cook, R.L. Stochastic sampling in computer graphics. ACM Trans. Graph. (TOG) 1986, 5, 5172. [Google Scholar] [CrossRef]
  29. Jung, H.; Ye, J.C.; Kim, E.Y. Improved k-t BLAST and k-t SENSE using FOCUSS. Phys. Med. Biol. 2007, 52, 3201–3226. [Google Scholar] [CrossRef]
  30. He, Z.; Cichocki, A.; Zdunek, R.; Xie, S. Improved FOCUSS method with conjugate gradient iterations. IEEE Trans. Signal Process. 2009, 57, 399–404. [Google Scholar]
  31. Deshmane, A.; Gulani, V.; Griswold, M.A.; Seiberlich, N. Parallel MR imaging. J. Magn. Reson. Imaging 2012, 36, 55–72. [Google Scholar] [CrossRef] [Green Version]
  32. Feng, L.; Xu, J.; Kim, D.; Axel, L.; Sodickson, D.K.; Otazo, R. Combination of compressed sensing, parallel imaging and partial Fourier for highly-accelerated 3D first-pass cardiac perfusion MRI. In Proceedings of the 19th International Society for Magnetic Resonance in Medicine (ISMRM), Quebec, QC, Canada, 7–13 May 2011; p. 4368. [Google Scholar]
  33. Margosian, P.; Schmitt, F.; Purdy, D. Faster MR imaging: Imaging with half the data. Healthc. Instrum. 1986, 1, 195–197. [Google Scholar]
  34. Cuppen, J.; van Est, A. Reducing MR imaging time by one-sided reconstruction. Magn. Reson. Med. 1987, 5, 526–527. [Google Scholar] [CrossRef]
  35. Otazo, R.; Kim, D.; Axel, L.; Sodickson, D. Combination of compressed sensing and parallel imaging for highly accelerated first-pass cardiac perfusion MRI. Magn. Reson. Med. 2010, 64, 767–776. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Liang, D.; Liu, B.; Wang, J.; Ying, L. Accelerating SENSE using compressed sensing. Magn. Reson. Med. 2009, 62, 1574–1584. [Google Scholar] [CrossRef] [PubMed]
  37. King, K.F.; Angelos, L. SENSE image quality improvement using matrix regularization. Int. Soc. Magn. Reson. Med. 2001, 9, 1771. [Google Scholar]
  38. Thüring, T.; Eggers, H.; Doneva, M.; Kozerke, S. A fast Reconstruction method using Compressive Sensing and an additional phase constraint. In Proceedings of the 26th European Society for Magnetic Resonance in Medicine and Biology (ESMRMB), Antalya, Turkey, 1–3 October 2009; p. 31. [Google Scholar]
  39. Malczewski, K. Super-Resolution with compressively sensed MR/PET signals at its input. Inform. Med. Unlocked 2020, 18, 1–20. [Google Scholar] [CrossRef]
  40. Malczewski, K. Rapid Diffusion Weighted Imaging with Enhanced Resolution. Appl. Magn. Reson. 2020, 51, 221–239. [Google Scholar] [CrossRef]
  41. Tsai, C.M.; Nishimura, D.G. Reduced aliasing artefacts using variable density k-space sampling trajectories. Magn. Reson. Med. 2000, 43, 452–458. [Google Scholar] [CrossRef]
  42. Lucia, M.; Granata, G.; Ichcha, M.; Mario, M.; Mario, R. GuarracinoGlioma Grade Classification via Omics Imaging. In Proceedings of the 7th International Conference on Bioimaging, Vienna, Austria, 24–26 February 2020. [Google Scholar]
  43. Heinrich, M.P.; Jenkinson, M.; Brady, M.; Schnabe, J.A. Globally optimal deformable registration on a minimum spanning tree using dense displacement sampling. In International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2012); Ayache, N., Delingette, H., Golland, P., Mori, K., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Part III. [Google Scholar]
  44. Raj, A.; Singh, G.; Zabih, R.; Kressler, B.; Wang, Y.; Schuff, N.; Weiner, M. Bayesian parallel imaging with edge-preserving priors. Magn. Reson. Med. 2007, 57, 8–21. [Google Scholar] [CrossRef] [Green Version]
  45. Yang, F.; Ding, M.; Zhang, X. Non-Rigid Multi-Modal 3D Medical Image Registration Based on Foveated Modality Independent Neighbourhood Descriptor. Sensors 2019, 19, 4675. [Google Scholar] [CrossRef] [Green Version]
  46. Jaya, V.L.; Gopikakumari, R. IEM: A New Image Enhancement Metric for Contrast and Sharpness Measurements. Int. J. Comput. Appl. 2013, 79, 1–9. [Google Scholar]

Share and Cite

MDPI and ACS Style

Malczewski, K. Fixing Acceleration and Image Resolution Issues of Nuclear Magnetic Resonance. Symmetry 2020, 12, 681. https://doi.org/10.3390/sym12040681

AMA Style

Malczewski K. Fixing Acceleration and Image Resolution Issues of Nuclear Magnetic Resonance. Symmetry. 2020; 12(4):681. https://doi.org/10.3390/sym12040681

Chicago/Turabian Style

Malczewski, Krzysztof. 2020. "Fixing Acceleration and Image Resolution Issues of Nuclear Magnetic Resonance" Symmetry 12, no. 4: 681. https://doi.org/10.3390/sym12040681

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop