Next Article in Journal
Glucose Biosensor Based on Glucose Oxidase Immobilized on BSA Cross-Linked Nanocomposite Modified Glassy Carbon Electrode
Next Article in Special Issue
YOLO Series for Human Hand Action Detection and Classification from Egocentric Videos
Previous Article in Journal
Rethinking LEO Mega-Constellation Routing to Provide Fast Internet Access Services
Previous Article in Special Issue
A Study on the Effectiveness of Deep Learning-Based Anomaly Detection Methods for Breast Ultrasonography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Successive Next Network as Augmented Regularization for Deformable Brain MR Image Registration

School of Information Science and Engineering, Linyi University, Linyi 276000, China
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(6), 3208; https://doi.org/10.3390/s23063208
Submission received: 5 February 2023 / Revised: 27 February 2023 / Accepted: 8 March 2023 / Published: 17 March 2023

Abstract

:
Deep-learning-based registration methods can not only save time but also automatically extract deep features from images. In order to obtain better registration performance, many scholars use cascade networks to realize a coarse-to-fine registration progress. However, such cascade networks will increase network parameters by an n-times multiplication factor and entail long training and testing stages. In this paper, we only use a cascade network in the training stage. Unlike others, the role of the second network is to improve the registration performance of the first network and function as an augmented regularization term in the whole process. In the training stage, the mean squared error loss function between the dense deformation field (DDF) with which the second network has been trained and the zero field is added to constrain the learned DDF such that it tends to 0 at each position and to compel the first network to conceive of a better deformation field and improve the network’s registration performance. In the testing stage, only the first network is used to estimate a better DDF; the second network is not used again. The advantages of this kind of design are reflected in two aspects: (1) it retains the good registration performance of the cascade network; (2) it retains the time efficiency of the single network in the testing stage. The experimental results show that the proposed method effectively improves the network’s registration performance compared to other state-of-the-art methods.

1. Introduction

Image registration is one of the basic tasks in medical image processing. It involves the acquisition of a dense deformation field (DDF) when a moving image is matched with a fixed image so that the two to-be-aligned images and their corresponding anatomical structures are aligned accurately in space [1]. The traditional registration method optimizes the cost function through a large number of iterations, a process that usually requires a significant amount of computation and time [2]. With the popularization and application of deep learning in the field of medical image registration, the deep learning registration method is now faster than the traditional image registration method. Therefore, for moving and fixed images, deformation fields can be generated by training a neural network, thus achieving rapid registration for a forward pass in the testing stage. Fan et al. [3] studied the computational costs of seven different deformable registration algorithms. The results showed that the assessed deep-learning network (BIRNet) without any iterative optimization needed the least time. Additionally, the registration accuracy improved after applying the deep learning method. For example, Cao et al. [4] proposed a deep learning method for registering brain MRI images, and it was revealed that the method’s Dice coefficient was improved in terms of registering white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF).
The unsupervised learning image registration method has been widely applied because it is not difficult to obtain gold-standard registration [5]. Balakrishnan et al. [6] optimized the U-Net neural network by defining the loss function as a combination of the mean square error similarity measure and the deformation field’s smoothing constraint. de Vos et al. [7] accomplished affine and deformable registration by superimposing several networks through unsupervised training. Kim et al. [8,9] used cyclic consistency to provide implicit regularization for maintaining topology and realizing 2D or 3D image registration. Moreover, a multi-scale strategy was adopted during the experiment to solve the relevant storage problem. Jiang et al. [10] proposed an unsupervised network framework (MJ-CNN) that adopted a multi-scale joint training scheme to achieve end-to-end optimization. Kong et al. [11] designed a cascade-connected channel attention mechanism network. During cascade registration, the attention module is incorporated to learn the features of the input image, thereby improving the expression ability of the image features. Through five iterations of the deformation field, improved bidirectional image registration was realized. Yang et al. [12] used multiple cascaded U-Net models to form a network structure. In their structure, each U-Net is trained with smooth regularization parameters to improve the accuracy of 3D medical image registration. Zhu et al. [13] helped a network develop high-similarity spatial correspondence by introducing a local attention model and integrated multi-scale functionality into the attention mechanism module to achieve the coarse-to-fine registration of local information. Ouyang et al. [14] trained their designed subnetworks synergistically by training the residual recursive cascade network to realize cooperation between the subnetworks. Through the connection of the residual network, the registration speed was accelerated. Guo et al. [15] improved the image registration accuracy and efficiency of CT-MR and used two cyclic consistency methods in a full convolution neural network to generate the spatial deformation field. Sideri-Lampretsa et al. [16] considered that it was easy to obtain edge images, so they used the image’s edges to drive the multimodal registration training process and thus help the network learn more effective information. Qian et al. [17] proposed a cascade framework of a registration network, and then registered images in training stages. The authors compared the performance of the cascade network framework with the traditional registration methods, subsequently, it was determined that the registration efficiency of the proposed method was significantly improved. Golkar et al. [18] proposed a hybrid registration framework of vessel extraction and thinning for retinal image segmentation, which improved the registration accuracy of complex retinal vessels.
Inspired by the idea of two-person zero-sum game from game theory, Goodefellow et al. [19] proposed a generation adversarial network (GAN) that used two neural networks for adversarial training and continuously improved the performance of the network in all directions during a game between the two networks. In addition to the in-depth study of the generative adversarial network (GAN), the application of an adversarial network has been integrated with techniques and aims from other fields, for instance, the combination of GAN and image processing. Therefore, GANs are also widely used in image registration. Santarossa et al. [20] used generation adversarial networks combined with ranking loss for multimodal image registration. Fan et al. [21,22] implemented a GAN in the unsupervised deformable registration of 3D brain MR images. In this approach, the discrimination network identifies whether a pair of images are sufficiently similar. The resulting feedback is then used to train the registration network. Simultaneously, GANs have been applied to single- and multi-mode image registration. Zheng et al. [23] used a GAN network to realize symmetric image registration and then transformed the symmetric registration formula of single- and multi-mode images into a conditional GAN. To align a pair of single-mode images, the registration method constitutes a cyclical process of transformation from one image to another and its inverse transformation. To align images with different modes, mode conversion should be performed before registration. In the training process, the method also adopts the semi-supervised method and trains using labeled and unlabeled images. Many registration methods have been produced based on the application of generation adversarial networks [24,25,26,27,28]. Huang et al. [29] fused a difficulty perception model into a cascade neural network composed of three networks. These networks are used to predict the coarse deformation field and the fine deformation field, respectively, so as to achieve accurate registration. GANs showed excellent performance in the aforementioned studies. In the previous study, a GAN based on dual attention mechanisms was proposed, which showed good registration performance in areas with relatively flat edges, but poor registration performance in narrow and long-edge areas. To this end, based on previous research, this paper proposes a method to assist GANs in realizing the registration of long and narrow regions at the peripheries of the brain, which differs from the methods of coarse registration and fine registration. Our main contributions are summarized as follows:
  • During training, the cascade networks are trained simultaneously to save network training time.
  • The second network is used as a loss function. The mean square error loss function added to the second network can constrain the deformation field output by the second network such that it tends to 0. Only the first network is used during testing, which saves testing time.
  • Coupled with the adversarial training of GANs, the registration performance of the first network is further improved.
The rest of this paper is organized as follows. Section 2 introduces the networks proposed in this paper in detail. Section 3 introduces the experimental datasets and evaluation indicators. Section 4 introduces the experimental results obtained from the HBN and ABIDE datasets. In Section 5, we provide a discussion. Finally, the conclusions are given in Section 6.

2. Methodology

This paper proposes a method combining adversarial learning with cascade learning. Joint training of cascaded networks can allow them to predict more accurate deformation fields. The first (registration) network is used to study the deformation field ϕ 1 . The second (augmented) network enables the first network to learn more deformations. A discrimination network improves the first network’s performance through adversarial training. The structures of each cascading network are similar to those of VoxelMorph [6]. The proposed overall learning framework is illustrated in Figure 1.

2.1. First (Registration) Network

The registration network is the first network in cascading framework. Its inputs are the fixed image F and the moving image M . Its output is the deformation field ϕ 1 , i.e., ϕ 1 = G ( F , M ) . This network realizes the alignment from M to F , i.e., F = M ( ϕ 1 ) , where M ( ϕ 1 ) is the warped image. Subsequently, the loss function between M ( ϕ 1 ) and F is calculated to drive the training process. This loss function includes three parts: intensity similarity loss Lsim, adversarial loss Ladv, and smooth regularization term Lsmooth.
The adversarial loss function of the registration network is:
L adv p = log 1 p , c P + log p , c P
where p is the output value of the discrimination network and c indicates the registration network input.
Local cross-correlation metric is used to calculate the similarity of the intensity between fixed image F and warped image M ( ϕ 1 ) . The specific formula of the loss function is:
C C F , M ϕ 1 = p Ω ( p i ( F p i F ( p ) ) ( M ϕ p i M ( ϕ p ) ) ) 2 ( p i F p i F p 2 ) ( p i ( M ϕ p i M ( ϕ ( p ) ) ) 2 )
where p i denotes the iteration of the n 3 volume center at voxel p , and Ω represents a three-dimensional voxel. In this paper, n = 9   F p i , and M ( ϕ 1 ( p i ) ) represents the voxel intensities of F and M ( ϕ 1 ) at p i , respectively. F ( p ) and M ( ϕ 1 ( p ) ) are the local mean values of n 3 volume. A higher CC indicates a more accurate alignment. According to the definition of CC, the intensity similarity loss Lsim is defined as follows:
L sim F , M ϕ 1 = C C ( F , M ( ϕ 1 ) )
Additionally, L2 regularization is implemented to smooth the deformation field ϕ 1 :
L smooth ϕ 1 = p Ω ϕ 1 ( p ) 2

2.2. Successive (Augmented) Network

The inputs of the successive network are F and M ( ϕ 1 ) ; the output is DDF ϕ 2 . ϕ 2 is used to deform M ( ϕ 1 ) to obtain ϕ 2 ( M ( ϕ 1 ) ) . Simultaneously, to clarify the warped image, we perform a composed operation on ϕ 1 and ϕ 2 , i.e., ϕ 1 ° ϕ 2 . M ( ϕ 1 ° ϕ 2 ) is obtained by the moving image M with the composed DDF. Next, two intensity loss functions, namely, Lsim( F , M ( ϕ 1 ° ϕ 2 ) ) and Lsim( F , ϕ 2 ( M ( ϕ 1 ) ) ), are calculated between M ( ϕ 1 ° ϕ 2 ) and F and between ϕ 2 ( M ( ϕ 1 ) ) and F , respectively. The DDF ϕ 2 is also constrained as it approaches zero deformation field through the following MSE loss function, allowing the deformation field ϕ 1 to learn more accurate deformations.
The formula of MSE loss function is defined as:
L mse ϕ 2 =   L mse ϕ 2 , 0 = p Ω ϕ 2 ( p ) 2
Through this function, the output effect of the first network can achieve fine registration after the two networks are connected in series.
The loss function for the registration network is as follows:
L G = L adv   ( p ) + α L sim ( F , M ( ϕ 1 ) ) + λ L smooth ( ϕ 1 )
In addition, the loss function used by the second network is:
L A = L sim F , M ϕ 1 + L sim ( M ( ϕ 1 ° ϕ 2 ) ) + L sim F , ϕ 2 M ϕ 1 +   L mse   ϕ 2 + L smooth   ϕ 1 + L smooth   ϕ 2 + L smooth ( ϕ 1 ° ϕ 2 )
The total loss function is:
L total = L G   +   L A

2.3. Discrimination Network

The discrimination network consists of four convolutional layers combined with leakyReLU activation layers. Finally, the sigmoid activation function is used to output the probability value. The discrimination network is shown in Figure 2. The discrimination network distinguishes the authenticity of image. The harder it is to distinguish the warped image from the fixed image, the harder it is to judge the authenticity of the image by the discrimination network.

3. Experiment

3.1. Experimental Details

Python and TensorFlow were used to implement the experimental process. The program was trained and tested with GPU NVIDIA GeForce GTX 2080 Ti [30].
In the training process, the patch-based training method is adopted to reduce the occupied memory. Herein, 127 blocks are obtained from each image with a size of 182 × 218 × 182. Each block size is 64 × 64 × 64. The stride is 32. The learning rates for training the registration and discrimination networks are set to 0.00001 and 0.000001, respectively.
The traditional methods of Demons and SyN are used as comparative experiments. The deep learning model VoxelMorph is also trained. VoxelMorph is a model of medical image registration based on unsupervised learning. Therefore, VoxelMorph is selected as the comparative experiment for deep learning. The Dice score, structural similarity, and Pearson’s correlation coefficient are used as the evaluation indicators to verify the superiority of the experimental results. Moreover, the influence of the MSE and Lsim loss functions on the experimental results is investigated.

3.2. Datasets

To prove the flexibility and superior performance of the proposed method, the HBN [31] and ABIDE datasets [32] are used for training and testing. The HBN dataset consists of brain data obtained from patients with ADHD (aged 5–21 years). Herein, 496 and 31 T1-weighted brain images are selected for training and testing, respectively. ABIDE is a dataset consisting of brain images from patients with autism (aged 5–64 years). Herein, 928 and 60 T1-weighted brain images are used for training and testing, respectively. The fixed image used in training comprises a pair of images randomly selected from the training set such that each image is linearly aligned to the fixed image. The image size of both the HBN and ABIDE datasets is 182 × 218 × 182 voxels with a resolution of 1 × 1 × 1 mm3. Both these datasets contain segmentation marker images of CSF, GM, and WM.

3.3. Evaluation Indicators

3.3.1. Dice Score

The Dice coefficient (Dice) index is used to evaluate the degree of overlap between a warped segmentation image and the segmentation image of the fixed image. This index reflects the similarity between the experimental and the standard segmentation images. It is defined as follows:
D i c e = 2 X s e g Y s e g X s e g Y s e g
where X s e g and Y s e g represent the standard and warped segmentation images, respectively. The range of Dice values is 0–1, corresponding to a range in the gap between the warped and the standard segmentation images progressing from large to small values, respectively. Alternatively, the closer the experimental result is to 1, the more similar the warped segmentation image is to the standard segmentation image, and the better is the registration result.

3.3.2. Structural Similarity

The structure similarity index measure [33] can measure the similarity of two images. The SSIM is calculated as:
SSIM X , Y = ( 2 μ X μ Y + c 1 ) ( 2 σ X Y + c 2 ) ( μ X 2 + μ Y 2 + c 1 ) ( σ X 2 + σ Y 2 + c 2 )
where X , Y represent the two input 3D images; μ X and μ Y represent the average value of X and Y , respectively. σ X 2 and σ Y 2 are the variances of X and Y , respectively. σ X and σ Y represent the standard deviation of X and Y , respectively. σ X Y represents the covariance of X and Y . c 1 and c 2 are constants used to avoid system errors caused by a denominator equal to 0. The SSIM can measure the structural similarity between the real and warped images. A SSIM value close to 1 indicates that the two images have a high degree of similarity.

3.3.3. Pearson’s Correlation Coefficient

Pearson’s correlation coefficient (PCC) was used to measure the similarity between two 3D images. The calculation formula of PCC is:
ρ ( X , Y ) i = 1 n ( X i X ) ( Y i Y ) i = 1 n ( X i X ) 2 i = 1 n ( Y i Y ) 2
The closer the value of PCC is to 1, the greater is the correlation. A PCC of 0 indicates no correlation. X , Y refer to the two input 3D images. X and Y represent the mean value of X and Y , respectively.

4. Results

The proposed methodology is compared with the following approaches: (1) Demons and SyN, two traditional registration methods; (2) Voxelmorph (VM), an unsupervised deep learning registration method; and (3) VM + A, a method consisting of a simultaneously trained registration network and augmented network.
First, the proposed GAN method (VM + A + GAN) is compared with Demons and SyN, which are two traditional methods. Table 1 and Table 2 summarize the test results obtained through different datasets, and all indicators show that our experimental results are the best. Figure 3 shows the comparison of the test results of the two datasets. The first row of the experimental image represents the original image obtained from the HBN dataset, and the second row represents the segmentation image corresponding to the original image derived from the HBN dataset. Similarly, the third row represents the original image based on the ABIDE dataset, and the fourth row represents the segmentation image corresponding to the original image derived from the ABIDE dataset. Compared with Demons and SyN, the image obtained by the proposed GAN method is closer in appearance to the fixed image, and the parts with differences are shown in the enlarged image on the right.
Second, the proposed GAN method is compared with the VM and VM + A methods. Figure 4 shows the registered moving image and the fixed image. Moreover, the first row represents the original image from the HBN dataset, and the second row represents the segmentation image corresponding to the original image from the HBN dataset. Similarly, the third row represents the original image from the ABIDE dataset, and the fourth row represents the segmentation image corresponding to the original image from the ABIDE dataset. Additionally, the enlarged figure on the right shows that the result for the proposed method regarding the training of the registration, augmented, and discrimination networks together is closer to the fixed image. Through the experimental results, the performance of the registration, augmented, and discrimination networks when trained together is verifiably better than that of the registration network trained individually and of the registration and augmented networks trained simultaneously.
In order to more clearly highlight the effectiveness of the method proposed in this paper, Figure 5 shows the experimental results of the three parts of the brain tissue based on the HBN dataset, and Figure 6 shows the experimental results of the three parts of the brain tissue based on the ABIDE dataset. The dotted circle in the figure is the result obtained by the method proposed in this paper.
Table 3 and Table 4 summarize the Dice, SSIM, and PCC indices corresponding to the different datasets. Considering Table 3, for the HBN dataset, the proposed method improves the precision values by 0.030, 0.032, and 0.034 compared with the VM method. For the ABIDE dataset, the proposed method improves the accuracies by 0.008, 0.004, and 0.004 compared with the VM method. Considering Table 4, for the HBN dataset, the proposed method increases the SSIM and PCC indices by 0.02 and 0.008, respectively, compared with the VM method. For the ABIDE dataset, the proposed method improves the SSIM and PCC indices by 0.006 and 0.003, respectively, compared with the VM method.

5. Discussion

The usage of a registration and discrimination networks for image registration is a common method. Such a registration method has been investigated experimentally in previous work [34]. However, this adversarial method for training a GAN only limitedly improves a registration network’s performance, and the registration capacity in some narrow and long edge areas needs to be further improved. Therefore, this paper proposes a method of training three networks together to allow the registration network to learn more deformations, further improving the registration performance. When the three networks are trained together, the use of different loss functions has a certain impact on the experimental results, which is discussed in the following subsections.

5.1. Importance of MSE

When two networks (VM + A) were trained together, both the Lsmooth loss function of the deformation field ϕ 2 and the MSE loss function were calculated. An experiment was also performed without the MSE loss function (VM + A − MSE) to verify its effectiveness. Additionally, when the three networks (VM + A + GAN) were trained together, the MSE loss function was removed again (VM + A + GAN − MSE), and experiments were performed to verify the impact of the MSE loss function on the experimental results. Through comparison, the best registration effect was achieved when the three networks were trained together and combined with the MSE loss function. The results are shown in Figure 7.
Table 5 summarizes the experimental results regarding the removal of the MSE loss function (VM + A − MSE) when two networks were trained together (VM + A) and the removal of the MSE loss function (VM + A + GAN − MSE) when three networks were trained together (VM + A + GAN). When comparing the results, note that the removal of the MSE loss function reduces registration accuracy, thus verifying that registration performance can be improved by adding the MSE loss function when these three networks are trained together. Comparing the SSIM and PCC metrics in Table 6, the loss function used by the proposed method achieves good results. Figure 4 shows the comparison of the experimental results after the MSE loss function was removed (VM + A − MSE) when two networks were trained together and after the MSE loss function was removed (VM + A + GAN − MSE) when three networks were trained together. Evidently, the proposed method obtained a result that is closer to the fixed image, which confirms the effectiveness of training three networks simultaneously; moreover, note that the proposed method intuitively shows a good registration effect in the narrow and long regions of the peripheries of the brain images. The first row of the resulting images represents the original image from the experimental results for the HBN dataset, and the second row represents the segmentation image corresponding to the original image from the experimental results for the HBN dataset. Similarly, the third row represents the original image from the experimental results for the ABIDE dataset, and the second row represents the segmentation image corresponding to the original image from the experimental results for the ABIDE dataset.

5.2. Importance of Lsim

When the three networks (VM + A + GAN) are trained together, the Lsmooth loss functions between the ϕ 2 ( M ( ϕ 1 ) ) image and the fixed image F as well as the M ( ϕ 1 ° ϕ 2 ) image and the fixed image F are removed for experimental comparison. After removing the two Lsim loss functions, the registration accuracy decreases significantly. Through this experimental analysis, it is evident that the Lsim loss function can restrict the similarity among the images to a certain extent, which proves the effectiveness of adding the Lsim loss function. By observing the histogram in Figure 8, it is evident that the proposed method improves the Dice, SSIM, and PCC indices. In Figure 8, note that (a) shows the importance of verifying the Lsim loss function for the HBN dataset; (b) shows the difference between verifying the proposed method for the ABIDE dataset and removing the Lsim loss function in the Dice index; (c) shows the impact of removing the Lsim loss function on the SSIM and PCC indices for the HBN dataset; and (d) shows the impact of removing the Lsim loss function on the SSIM and PCC indices for the ABIDE dataset.

5.3. Importance of Different Deformation Fields

The Dice values for when two networks were trained simultaneously are calculated and discussed next to verify ϕ 1 , ϕ 2 , and ϕ 1 ° ϕ 2 in the images.
For ϕ 1 , the similarity is calculated between the warped moving image segmentation image M s e g ( ϕ 1 ) and the fixed image segmentation image F s e g , expressed as M s e g ϕ 1 F s e g . For ϕ 2 , the similarity is calculated between the warped ( M s e g ϕ 1 ) ( ϕ 2 ) and the fixed image segmentation image F s e g , expressed as M s e g ϕ 1 ϕ 2 F s e g . For ϕ 1 ° ϕ 2 , the similarity is calculated between the warped moving image segmentation image M s e g ϕ 1 ° ϕ 2 and the fixed image segmentation image F s e g , expressed as M s e g ϕ 1 ° ϕ 2 F s e g .
Considering the Dice values in the Table 7, the deformation field ( ϕ 2 ) still plays a certain role in image registration, but a significantly miniscule role. Therefore, the registration network still allows the deformation field ( ϕ 1 ) to learn more deformations, and the augmented network only plays a secondary role.

6. Conclusions

In this paper, a method wherein three networks (registration, augmented, and discrimination networks) are trained together is proposed, for which the MSE loss function is introduced into the augmented network to improve the registration network’s performance. It was demonstrated that the registration network’s performance was further improved when coupled with the adversarial capacity of a GAN. Then, it was proven that the proposed method offers significant advantages over the existing methods. In addition, it was clarified that the proposed training method is easy to implement, and that the implemented loss function is easy to obtain.
In the future, a more novel GAN will be used to further improve image registration performance; moreover, more indicators will be used for comparison. The developed model will then be tested on different datasets to prove its excellent generalizability.

Author Contributions

Conceptualization, writing—original draft preparation, writing—review and editing, M.L. (Meng Li); methodology, S.H.; software, G.L.; validation, M.L. (Mingtao Liu), F.Z. and J.L.; formal analysis, S.H.; investigation, Y.Y.; resources, L.Z.; data curation, Y.X.; visualization, Y.Y.; supervision, D.F.; project administration, W.Z.; funding acquisition, X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under Grant 61771230 and Shandong Provincial Natural Science Foundation: ZR2021MF115.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

All the datasets used to train the model presented in this paper were obtained from the Internet.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hu, Y.; Modat, M.; Gibson, E.; Li, W.; Ghavami, N.; Bonmati, E.; Wang, G.; Bandula, S.; Moore, C.M.; Mberton, M.; et al. Weakly-supervised convolutional neural networks for multimodal image registration. Med. Image Anal. 2018, 49, 1–13. [Google Scholar] [CrossRef]
  2. Shan, S.; Yan, W.; Guo, X.; Chang, E.I.; Fan, Y.; Xu, Y. Unsupervised end-to-end learning for deformable medical image registration. arXiv 2017, arXiv:1711.08608. [Google Scholar]
  3. Fan, J.; Cao, X.; Yap, P.T.; Shen, D. BIRNet: Brain image registration using dual-supervised fully convolutional networks. Med. Image Anal. 2019, 54, 193–206. [Google Scholar] [CrossRef]
  4. Cao, X.; Yang, J.; Zhang, J.; Wang, Q.; Yap, P.T.; Shen, D. Deformable image registration using a cue-aware deep regression network. IEEE Trans. Biomed. Eng. 2018, 65, 1900–1911. [Google Scholar] [CrossRef]
  5. Wang, C.; Yang, G.; Papanastasiou, G. Unsupervised image registration towards enhancing performance and explainability in cardiac and brain image analysis. Sensors 2022, 22, 2125. [Google Scholar] [CrossRef]
  6. Balakrishnan, G.; Zhao, A.; Sabuncu, M.R.; Guttag, J.; Dalca, A.V. VoxelMorph: A learning framework for deformable medical image registration. IEEE Trans. Med. Imaging 2019, 38, 1788–1800. [Google Scholar] [CrossRef] [Green Version]
  7. de Vos, B.D.; Berendsen, F.F.; Viergever, M.A.; Sokooti, H.; Staring, M.; Išgum, I. A deep learning framework for unsupervised affine and deformable image registration. Med. Image Anal. 2019, 52, 128–143. [Google Scholar] [CrossRef] [Green Version]
  8. Kim, B.; Kim, D.H.; Park, S.H.; Kim, J.; Lee, J.G.; Ye, J.C. CycleMorph: Cycle consistent unsupervised deformable image registration. Med. Image Anal. 2021, 71, 102036. [Google Scholar] [CrossRef]
  9. Kim, B.; Kim, J.; Lee, J.G.; Kim, D.H.; Park, S.H.; Ye, J.C. Unsupervised deformable image registration using cycle-consistent cnn. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 166–174. [Google Scholar]
  10. Jiang, Z.; Yin, F.F.; Ge, Y.; Ren, L. A multi-scale framework with unsupervised joint training of convolutional neural networks for pulmonary deformable image registration. Phys. Med. Biol. 2020, 65, 015011. [Google Scholar] [CrossRef]
  11. Kong, L.; Yang, T.; Xie, L.; Xu, D.; He, K. Cascade connection-based channel attention network for bidirectional medical image registration. Vis. Comput. 2022, 1–19. [Google Scholar] [CrossRef]
  12. Yang, J.; Wu, Y.; Zhang, D.; Cui, W.; Yue, X.; Du, S.; Zhang, H. LDVoxelMorph: A precise loss function and cascaded architecture for unsupervised diffeomorphic large displacement registration. Med. Phys. 2022, 49, 2427–2441. [Google Scholar] [CrossRef]
  13. Zhu, F.; Wang, S.; Li, D.; Li, Q. Similarity attention-based CNN for robust 3D medical image registration. Biomed. Signal Process. Control. 2023, 81, 104403. [Google Scholar] [CrossRef]
  14. Ouyang, X.; Liang, X.; Xie, Y. Preliminary feasibility study of imaging registration between supine and prone breast CT in breast cancer radiotherapy using residual recursive cascaded networks. IEEE Access 2020, 9, 3315–3325. [Google Scholar] [CrossRef]
  15. Guo, Y.; Wu, X.; Wang, Z.; Pei, X.; Xu, X.G. End-to-end unsupervised cycle-consistent fully convolutional network for 3D pelvic CT-MR deformable registration. J. Appl. Clin. Med. Phys. 2020, 21, 193–200. [Google Scholar] [CrossRef]
  16. Sideri-Lampretsa, V.; Kaissis, G.; Rueckert, D. Multi-modal unsupervised brain image registration using edge maps. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  17. Qian, L.; Zhou, Q.; Cao, X.; Shen, W.; Suo, S.; Ma, S.; Qu, G.; Gong, X.; Yan, Y.; Jiang, L.; et al. A cascade-network framework for integrated registration of liver DCE-MR images. Comput. Med. Imaging Graph. 2021, 89, 101887. [Google Scholar] [CrossRef]
  18. Golkar, E.; Rabbani, H.; Dehghani, A. Hybrid Registration of Retinal Fluorescein Angiography and Optical Coherence Tomography Images of Patients with Diabetic Retinopathy. Biomed. Opt. Express 2021, 12, 1707–1724. [Google Scholar] [CrossRef]
  19. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar] [CrossRef]
  20. Santarossa, M.; Kilic, A.; von der Burchard, C.; Schmarje, L.; Zelenka, C.; Reinhold, S.; Koch, R.; Roider, J. MedRegNet: Unsupervised multimodal retinal-image registration with GANs and ranking loss. In Medical Imaging 2022: Image Processing; SPIE: San Diego, CA, USA, 2022; Volume 12032, pp. 321–333. [Google Scholar]
  21. Fan, J.; Cao, X.; Xue, Z.; Yap, P.T.; Shen, D. Adversarial similarity network for evaluating image alignment in deep learning based registration. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 739–746. [Google Scholar]
  22. Fan, J.; Cao, X.; Wang, Q.; Yap, P.T.; Shen, D. Adversarial learning for mono-or multi-modal registration. Med. Image Anal. 2019, 58, 101545. [Google Scholar] [CrossRef]
  23. Zheng, Y.; Sui, X.; Jiang, Y.; Che, T.; Zhang, S.; Yang, J.; Li, H. SymReg-GAN: Symmetric image registration with generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5631–5646. [Google Scholar] [CrossRef]
  24. Yan, P.P.; Xu, S.; Rastinehad, A.R.; Wood, B.J. Adversarial image registration with application for MR and TRUS image fusion. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Granada, Spain, 16 September 2018; pp. 197–204. [Google Scholar]
  25. Duan, L.; Yuan, G.; Gong, L.; Fu, T.; Yang, X.; Chen, X.; Zheng, J. Adversarial learning for deformable registration of brain MR image using a multi-scale fully convolutional network. Biomed. Signal Proces. Control 2019, 53, 101562. [Google Scholar] [CrossRef]
  26. Tanner, C.; Ozdemir, F.; Profanter, R.; Vishnevsky, V.; Konukoglu, E.; Goksel, O. Generative adversarial networks for MR-CT deformable image registration. arXiv 2018, arXiv:1807.07349. [Google Scholar]
  27. Hu, Y.; Gibson, E.; Ghavami, N.; Bonmati, E.; Moore, C.M.; Emberton, M.; Vercauteren, T.; Noble, J.A.; Barrat, D.C. Adversarial deformation regularization for training image registration neural networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 774–782. [Google Scholar]
  28. Lei, Y.; Fu, Y.; Wang, T.; Liu, Y.; Patel, P.; Curran, W.J.; Liu, T.; Yang, X. 4D-CT deformable image registration using multiscale unsupervised deep learning. Phys. Med. Biol. 2020, 65, 085003. [Google Scholar] [CrossRef] [PubMed]
  29. Huang, Y.; Ahmad, S.; Fan, J.; Shen, D.; Yap, P.T. Difficulty-aware hierarchical convolutional neural networks for deformable registration of brain MR images. Med. Image Anal. 2021, 67, 101817. [Google Scholar] [CrossRef] [PubMed]
  30. Wu, L.; Hu, S.; Liu, C. Exponential-distance weights for reducing grid-like artifacts in patch-based medical image registration. Sensors 2021, 21, 7112. [Google Scholar] [CrossRef] [PubMed]
  31. Alexander, L.M.; Escalera, J.; Ai, L.; Andreotti, C.; Febre, K.; Mangone, A.; Vega-Potler, N.; Langer, N.; Alexander, A.; Kovacs, M.; et al. An open resource for transdiagnostic research in pediatric mental health and learning disorders. Sci. Data 2017, 4, 170181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Craddock, C.; Benhajali, Y.; Chu, C.; Chouinard, F.; Evans, A.; Jakab, A.; Khundrakpam, B.S.; Lewis, J.D.; Li, Q.; Milham, M.; et al. The Neuro Bureau Preprocessing Initiative: Open sharing of preprocessed neuroimaging data and derivatives. Front. Neuroinform. 2013, 7, 27. [Google Scholar]
  33. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  34. Li, M.; Wang, Y.; Zhang, F.; Li, G.; Hu, S.; Wu, L. Deformable medical image registration based on unsupervised generative adversarial network integrating dual attention mechanisms. In Proceedings of the 14th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 23–25 October 2021; pp. 1–6. [Google Scholar]
Figure 1. Overall network framework.
Figure 1. Overall network framework.
Sensors 23 03208 g001
Figure 2. The overall framework of the adversarial network. The adversarial network consists of convolution and LeakyReLU activation layer.
Figure 2. The overall framework of the adversarial network. The adversarial network consists of convolution and LeakyReLU activation layer.
Sensors 23 03208 g002
Figure 3. Registration results of Demons, SyN, and our proposed method using HBN and ABIDE datasets.
Figure 3. Registration results of Demons, SyN, and our proposed method using HBN and ABIDE datasets.
Sensors 23 03208 g003
Figure 4. Registration results based on deep learning methods. Among them, VM represents the result obtained by the VoxelMorph method, VM + A represents the result obtained by training the registration network and the enhanced network together, and VM + A + GAN represents the result obtained by our method.
Figure 4. Registration results based on deep learning methods. Among them, VM represents the result obtained by the VoxelMorph method, VM + A represents the result obtained by training the registration network and the enhanced network together, and VM + A + GAN represents the result obtained by our method.
Sensors 23 03208 g004
Figure 5. GAN registration performance on the HBN dataset.
Figure 5. GAN registration performance on the HBN dataset.
Sensors 23 03208 g005
Figure 6. GAN registration performance on the ABIDE dataset.
Figure 6. GAN registration performance on the ABIDE dataset.
Sensors 23 03208 g006
Figure 7. Experimental results regarding the use of the MSE loss function when employing the HBN and ABIDE datasets. Among them, VM + A − MSE indicates that the MSE loss function has been removed when training the registration network and the enhanced network, VM + A indicates the experimental results when the MSE loss function is retained when training the registration network and the enhanced network, VM + A + GAN − MSE indicates our method’s experimental results following the removal of the MSE loss function, and VM + A + GAN represents our experimental results with the MSE loss function retained.
Figure 7. Experimental results regarding the use of the MSE loss function when employing the HBN and ABIDE datasets. Among them, VM + A − MSE indicates that the MSE loss function has been removed when training the registration network and the enhanced network, VM + A indicates the experimental results when the MSE loss function is retained when training the registration network and the enhanced network, VM + A + GAN − MSE indicates our method’s experimental results following the removal of the MSE loss function, and VM + A + GAN represents our experimental results with the MSE loss function retained.
Sensors 23 03208 g007
Figure 8. Influence of Lsim loss function on registration results.
Figure 8. Influence of Lsim loss function on registration results.
Sensors 23 03208 g008
Table 1. Dice values obtained with the HBN and ABIDE datasets. Bold numbers indicate the best results.
Table 1. Dice values obtained with the HBN and ABIDE datasets. Bold numbers indicate the best results.
HBNABIDE
MethodsCSFGMWMCSFGMWM
Demons0.513 ± 0.1580.335 ± 0.3200.345 ± 0.3300.410 ± 0.2280.312 ± 0.3050.332 ± 0.320
SyN0.585 ± 0.0260.768 ± 0.0220.786 ± 0.0150.593 ± 0.0410.749 ± 0.0190.791 ± 0.023
VM + A + GAN0.653 ± 0.0410.829 ± 0.0290.855 ± 0.0150.646 ± 0.0460.801 ± 0.0240.847 ± 0.019
Table 2. SSIM and PCC metrics obtained with the HBN and ABIDE datasets. Bold numbers indicate the best results.
Table 2. SSIM and PCC metrics obtained with the HBN and ABIDE datasets. Bold numbers indicate the best results.
HBNABIDE
MethodsSSIMPCCSSIMPCC
Demons0.7810.8860.7630.886
SyN0.9040.9620.8700.958
VM + A + GAN0.9560.9840.9200.985
Table 3. Dice indicator based on deep learning. Bold numbers indicate the best results.
Table 3. Dice indicator based on deep learning. Bold numbers indicate the best results.
HBNABIDE
MethodsCSFGMWMCSFGMWM
VM0.623 ± 0.0370.797 ± 0.0270.821 ± 0.0150.638 ± 0.0480.797 ± 0.0240.843 ± 0.018
VM + A0.642 ± 0.0420.821 ± 0.0290.846 ± 0.0140.639 ± 0.0480.796 ± 0.0230.841 ± 0.018
VM + A + GAN0.653 ± 0.0410.829 ± 0.0290.855 ± 0.0150.646 ± 0.0460.801 ± 0.0240.847 ± 0.019
Table 4. SSIM and PCC metrics for deep-learning-based registration methods. Bold numbers indicate the best results.
Table 4. SSIM and PCC metrics for deep-learning-based registration methods. Bold numbers indicate the best results.
HBNABIDE
MethodsSSIMPCCSSIMPCC
VM0.9360.9760.9140.982
VM + A0.9360.9760.9140.982
VM + A + GAN0.9560.9840.9200.985
Table 5. Dice values when using the MSE loss function and the HBN and ABIDE datasets. Bold numbers indicate the best results.
Table 5. Dice values when using the MSE loss function and the HBN and ABIDE datasets. Bold numbers indicate the best results.
MethodsHBNABIDE
CSFGMWMCSFGMWM
VM + A − MSE0.641 ± 0.0410.821 ± 0.0290.840 ± 0.0140.638 ± 0.0480.797 ± 0.0240.843 ± 0.018
VM + A0.642 ± 0.0420.821 ± 0.0290.846 ± 0.0140.639 ± 0.0480.796 ± 0.0230.841 ± 0.018
VM + A + GAN − MSE0.652 ± 0.0410.829 ± 0.0280.854 ± 0.0140.645 ± 0.0470.800 ± 0.0240.846 ± 0.019
VM + A + GAN0.653 ± 0.0410.829 ± 0.0290.855 ± 0.0150.646 ± 0.0460.801 ± 0.0240.847 ± 0.019
Table 6. SSIM and PCC values when using the MSE loss function and the HBN and ABIDE datasets. Bold numbers indicate the best results.
Table 6. SSIM and PCC values when using the MSE loss function and the HBN and ABIDE datasets. Bold numbers indicate the best results.
MethodsHBNABIDE
SSIMPCCSSIMPCC
VM + A − MSE0.9500.9820.9110.982
VM + A0.9520.9830.9100.981
VM + A + GAN − MSE0.9570.9850.9190.985
VM + A + GAN0.9560.9840.9200.985
Table 7. Test results of output images from registration and augmented network for two datasets. Bold numbers indicate the best results.
Table 7. Test results of output images from registration and augmented network for two datasets. Bold numbers indicate the best results.
MethodsHBNABIDE
CSFGMWMCSFGMWM
M s e g ϕ 1 F s e g 0.642 ± 0.0420.821 ± 0.0290.846 ± 0.0140.639 ± 0.0480.796 ± 0.0230.841 ± 0.018
M s e g ϕ 1 ϕ 2 F s e g 0.644 ± 0.0420.825 ± 0.0300.853 ± 0.0150.642 ± 0.0480.802 ± 0.0230.848 ± 0.018
M s e g ϕ 1 ° ϕ 2 F s e g 0.643 ± 0.0420.821 ± 0.0280.845 ± 0.0130.640 ± 0.0480.796 ± 0.0230.839 ± 0.018
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, M.; Hu, S.; Li, G.; Zhang, F.; Li, J.; Yang, Y.; Zhang, L.; Liu, M.; Xu, Y.; Fu, D.; et al. The Successive Next Network as Augmented Regularization for Deformable Brain MR Image Registration. Sensors 2023, 23, 3208. https://doi.org/10.3390/s23063208

AMA Style

Li M, Hu S, Li G, Zhang F, Li J, Yang Y, Zhang L, Liu M, Xu Y, Fu D, et al. The Successive Next Network as Augmented Regularization for Deformable Brain MR Image Registration. Sensors. 2023; 23(6):3208. https://doi.org/10.3390/s23063208

Chicago/Turabian Style

Li, Meng, Shunbo Hu, Guoqiang Li, Fuchun Zhang, Jitao Li, Yue Yang, Lintao Zhang, Mingtao Liu, Yan Xu, Deqian Fu, and et al. 2023. "The Successive Next Network as Augmented Regularization for Deformable Brain MR Image Registration" Sensors 23, no. 6: 3208. https://doi.org/10.3390/s23063208

APA Style

Li, M., Hu, S., Li, G., Zhang, F., Li, J., Yang, Y., Zhang, L., Liu, M., Xu, Y., Fu, D., Zhang, W., & Wang, X. (2023). The Successive Next Network as Augmented Regularization for Deformable Brain MR Image Registration. Sensors, 23(6), 3208. https://doi.org/10.3390/s23063208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop