Next Article in Journal
Social Media Opinion Analysis Model Based on Fusion of Text and Structural Features
Previous Article in Journal
Use of Underwater-Image Color to Determine Suspended-Sediment Concentrations Transported to Coastal Regions
Previous Article in Special Issue
Automated Classification of Urinary Cells: Using Convolutional Neural Network Pre-trained on Lung Cells
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of Artificial Intelligence-Based Dual-Energy Subtraction for Chest Radiography

1
Division of Health Sciences, Graduate School of Medicine, Osaka University, Suita 565-0871, Japan
2
Department of Radiology, National Cerebral and Cardiovascular Center, Suita 564-8565, Japan
3
Department of Radiology, Kitasato University Hospital, Sagamihara 252-0329, Japan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(12), 7220; https://doi.org/10.3390/app13127220
Submission received: 17 May 2023 / Revised: 9 June 2023 / Accepted: 15 June 2023 / Published: 16 June 2023

Abstract

:
Recently, some facilities have utilized the dual-energy subtraction (DES) technique for chest radiography to increase pulmonary lesion detectability. However, the availability of the technique is limited to certain facilities, in addition to other limitations, such as increased noise in high-energy images and motion artifacts with the one-shot and two-shot methods, respectively. The aim of this study was to develop artificial intelligence-based DES (AI–DES) technology for chest radiography to overcome these limitations. Using a trained pix2pix model on clinically acquired chest radiograph pairs, we successfully converted 130 kV images into virtual 60 kV images that closely resemble the real images. The averaged peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) between virtual and real 60 kV images were 33.8 dB and 0.984, respectively. We also achieved the production of soft-tissue- and bone-enhanced images using a weighted image subtraction process with the virtual 60 kV images. The soft-tissue-enhanced images exhibited sufficient bone suppression, particularly within lung fields. Although the bone-enhanced images contained artifacts on and around the lower thoracic and lumbar spines, superior sharpness and noise characteristics were presented. The main contribution of our development is its ability to provide selectively enhanced images for specific tissues using only high-energy images obtained via routine chest radiography. This suggests the potential to improve the detectability of pulmonary lesions while addressing challenges associated with the existing DES technique. However, further improvements are necessary to improve the image quality.

1. Introduction

Lung cancer is the disease with the highest mortality and the second-highest incidence of cancer worldwide [1,2]. Since early-stage lung cancer may have a better prognosis with appropriate treatment, early diagnosis and accurate staging are critical [2,3]. Randomized controlled trials, including the National Lung Screening Trial (NLST), have demonstrated that the use of low-dose computed tomography (CT) for lung cancer screening reduces the mortality by 20% compared to chest radiography [4,5]. Chest radiography lacks effectiveness [4], and the American Cancer Society Lung Cancer Screening Guidelines recommend low-dose CT rather than chest radiography [6]. Nevertheless, due to the low cost, low radiation dose, and high adoption rate of the equipment, chest radiography is widely performed for lung cancer screening, in addition to low-dose CT [7,8,9].
Additionally, some facilities utilize the dual-energy subtraction (DES) technique for chest radiography [10]. This technique can produce images that emphasize tissues with particular linear attenuation coefficients [11] and typically produces soft-tissue-enhanced and bone-enhanced images [12]. It has been reported that soft-tissue-enhanced images can improve the ability to detect pulmonary lesions [13,14,15,16]. Oda et al., compared the performance of radiologists in detecting pulmonary lesions by clinical chest radiography with and without the DES technique, and their receiver operating characteristic (ROC) analysis demonstrated the statistically significant superiority of using DES images [15]. Manji et al., reported that soft-tissue-enhanced images obtained through the DES technique statistically significantly reduced the reading time of radiologists and slightly improved the diagnostic accuracy of pulmonary lesions [16]. The superiority of soft-tissue-enhanced images obtained via the DES technique in the diagnosis of COVID-19 has been also confirmed [17]. Furthermore, research on advanced DES techniques, such as the optimization of exposure conditions [18] and the automatic determination of weight factors for bone and soft tissue enhancements in the subtraction process [19], has been actively reported.
However, such a DES technique has some problems. First, only limited numbers of facilities own the necessary systems. Second, there are several problems associated with the imaging techniques of the one-shot and two-shot methods. In the one-shot method, two images are simultaneously obtained at different energies by placing a thin copper plate between two imaging detectors [20]. Because of the copper plate, noise characteristics can deteriorate the quality of high-energy images [21]. In the two-shot method, on the other hand, X-ray exposure is carried out twice at different energies [22]. As a result, motion artifacts and dose increments are unavoidable.
Therefore, we aim to address these issues by developing a technique to virtually generate low-energy images from high-energy images using artificial intelligence (AI). This development of AI-based DES (AI–DES) does not require a specific imaging detector with a metal plate or multiple exposures. Moreover, the AI-DES can arbitrarily select the enhanced tissue by adjusting the weight of the image subtraction process. Consequently, our AI-DES has the potential to provide more enriched information than existing methods, where bone-suppressed images are directly produced [23,24,25,26,27,28,29,30,31]. For instance, Liu et al., developed an AI model to generate DES-like soft tissue images, but their approach primarily focused on suppressing bone tissues, making it difficult to selectively enhance specific tissues [25]. Similarly, Bae et al., developed a generative adversarial network (GAN)-based bone suppression model for chest radiography, and demonstrated that its ability to detect pulmonary lesions is comparable to that of a DES technique [26]. Cho et al., achieved bone suppression on pediatric chest radiographs by utilizing computed tomographic images of adults and pediatrics to train the AI model [27]. In contrast, our AI-DES, by adapting weighted image subtraction with artificially synthesized low-energy images, has a significant advantage in terms of generating not only soft-tissue- or bone-enhanced images but also selectively enhanced images of tissues with a specific linear attenuation coefficient. However, this paper mainly focuses on generating soft-tissue- and bone-enhanced images for comparison with existing DES systems as an initial report on the development of AI-DES.
We employed pix2pix [32], which is a well-established image-to-image translation network, to construct the AI network. Pix2pix has been widely used in many image domain transformation tasks, such as image colorization and style transfer. It is also extensively used for medical imaging. Yoshida et al., adopted pix2pix to correct motion artifacts in magnetic resonance (MR) images [33]. Sun et al., achieved denoising of low-dose single-photon emission-computed tomography (SPECT) images using pix2pix [34]. Although pix2pix usually requires identical positional information between the paired images, our task of generating low-energy images from high-energy input images can deal with this constraint.
The main contributions of this work are as follows:
  • We developed an AI-based DES system to provide soft-tissue- and bone-enhanced images using virtually generated low-energy images;
  • The virtual low-energy images were generated through the AI technique from only high-energy images, which can be obtained by routine chest radiography;
  • AI-DES has the potential to provide specific tissue-enhanced images while avoiding issues associated with DES systems, such as multiple exposures and noise increments;
  • A comparison of the generated images with those produced by a clinically applied system suggests that AI-DES can achieve superior sharpness and noise characteristics.
Furthermore, although this is a future-expanded perspective, the novelty of the AI-DES is that it allows for the selection of enhanced tissues by adjusting the weight in the image subtraction process, in comparison to existing works.
In Section 2, we first introduce the developed AI-DES system. Then, we describe image datasets and prepossessing steps for AI training. Next, we specify the training setups and explain the methods used to evaluate similarity between the generated and ground truth images. Section 3 presents the generated images in comparison to the ground truth. Section 4 discusses the performance, limitations, and future perspectives of our AI-DES. Finally, Section 5 summarizes and concludes this work.

2. Materials and Methods

2.1. AI-DES Development

Our developed AI-DES consists of an AI network and a weighted image subtraction process. We first describe the AI network, then explain the image subtraction process.

2.1.1. AI Network

We employed pix2pix for our AI network to convert high-energy images into low-energy images. This is a variant of a conditional generative adversarial network (cGAN) [35]. Unlike a typical cGAN that produces images from a random noise vector, the generator of pix2pix takes images as input and transforms them into images of a different domain by learning the relationship between the two domains [32]. The discriminator receives an image pair of two domains and attempts to determine whether the pair is real or fake.
Figure 1 illustrates the pix2pix network used in this study. The generator learns to convert high-energy images into virtual low-energy images similar to the corresponding real low-energy images. Simultaneously, the discriminator aims to distinguish between the pair of real high-energy and virtual low-energy images and the pair of real high-energy and low-energy images. The learning process is expressed as a min–max game with an adversarial loss function given by
min G max D L c G A N ( G , D ) = E x P high , y P low [ log D ( x , y ) ] + E x P high [ log ( 1 D ( x , G ( x ) ) ) ] ,
where x P h i g h represents high-energy images, y P l o w represents low-energy images, G is the generator, and D is the discriminator [34]. G attempts to minimize Equation (1), while D attempts to maximize it.
Pix2pix also imposes a constraint on the L1 distance between the generated and real images to make the generator produce images that are closer to the ground truth, as follows:
L 1 ( G ) = E x P high , y P low [ | | y G ( x ) | | 1 ] .
Thus, the objective function of pix2pix can be expressed as:
G * = min G max D L c G A N ( G , D ) + λ L 1 ( G ) ,
where λ = 100 was used in this study.
We implemented the pix2pix network by modifying a publicly available code [36]. Specifically, we changed the resolution of each layer of the generator and discriminator to produce images with a resolution of 1024 × 1024. Otherwise, the same architecture as the original pix2pix [32] was used. The details are described below and in Section 2.3.
Figure 2 and Table 1 show the generator network architecture in AI-DES. The generator has a 16-layer U-Net [37] structure with symmetric encoder and decoder parts. The encoder part has eight layers, consisting of two-dimensional convolution (Conv2d: kernel size = 4, stride = 2, padding = 1), batch normalization (BN), and a LekyReLU activation function. The decoder part also has eight layers, consisting of two-dimensional deconvolution (Deconv2d: kernel size = 4, stride = 2, padding = 1), BN, and a ReLU activation function. Dropout layers were inserted between the first and second, second and third, and third and fourth layers of the decoder part. The tanh activation function was applied to the final layer to output the virtual low-energy images. Most importantly, the U-Net architecture has skip connections that concatenate the mirrored encoder and decoder layers to recover high-frequency components.
The discriminator has five convolutional neural network (CNN) layers, as shown in Figure 3 and Table 2. The first layer takes six channels, since the discriminator receives a pair of two images, each with three channels. The first through third layers downsample the feature maps using Conv2d (kernel size = 4, stride = 2, padding = 1), while the fourth and final layers reduce the map resolution by one pixel using Conv2d (kernel size = 4, stride = 1, padding = 1). The discriminator employs the PatchGAN [38] approach to evaluate multiple image patches and averages the loss over the output map (126 × 126 × 1) to distinguish between real and virtual images.

2.1.2. Weighted Image Subtraction

We assumed that raw data of monochromatic low- and high-energy images in direct-conversion flat-panel detector (d-FPD) systems have pixel values of P L and P H , respectively as expressed by [11,39]
log 10 ( P L ) = ( μ B ( L ) · t B + μ S ( L ) · t S ) ,
log 10 ( P H ) = ( μ B ( H ) · t B + μ S ( H ) · t S ) ,
where μ B and μ S are the linear attenuation coefficients of bone and soft tissues at low (L) or high (H) energy, respectively; t B and t S denote the thicknesses of bone and soft tissues, respectively; and P L and P H are proportional to low- and high-energy X-ray intensity transmitted through the tissues, respectively, since d-FPD systems have a linear response to X-ray intensity. The weighted subtraction of Equations (4) and (5) is given by:
K H · log 10 ( P H ) K L · log 10 ( P L ) = ( K L · μ B ( L ) K H · μ B ( H ) ) t B + ( K L · μ S ( L ) K H · μ S ( H ) ) t S ,
where K L and K H are weight factors.
When ( K L · μ B ( L ) K H · μ B ( H ) ) equals zero, Equation (6) represents the emphasized difference in logarithmically amplified data of X-ray intensity transmitted through soft tissues at low and high energies. Finally, it corresponds to the pixel values ( P s u b ) of the specific tissue-enhanced images, as expressed by
P s u b = K H · log 10 ( P H ) K L · log 10 ( P L ) .
Apart from the raw data, such as P L and P H , P s u b denotes the pixel values for viewing, where log-amplified X-ray intensity distribution is exhibited.
Here, the weight factor ( ω S ) for soft-tissue-enhanced image generation is given by:
ω S = K H K L = μ B ( L ) μ B ( H ) .
Similarly, the weight factor ( ω B ) for bone-enhanced images is given by:
ω B = K H K L = μ S ( L ) μ S ( H ) .
It should be noted that ω S and ω B in Equations (8) and (9) represent theoretical values. In this study, we obtained the P s u b of soft-tissue- and bone-enhanced images by using the raw data of real high-energy and virtual low-energy images, as expressed by
P s u b = ω · log 10 ( P H ) log 10 ( P L ) ,
where ω is a weight factor. We adjusted the value of ω for each test case to most selectively emphasize the target tissues. We focused on generating soft-tissue- and bone-enhanced images to compare the performance with that of an existing DES system. However, the AI-DES can provide the option of arbitrarily targeting specific tissues for enhancement by adjusting the weight factor.

2.2. Dataset Preparation

We used raw data of chest radiographs taken by a clinically applied two-shot DES system (Discovery XR656, GE Healthcare, Chicago, IL, USA) at Kitasato University Hospital (Sagamihara City, Japan) to create our datasets. The tube voltages were 130 kV for high-energy images and 60 kV for low-energy images. The imaging detector, consisting of amorphous silicon with a cesium iodide (CSI) scintillator, is a d-FPD type with 3524 × 4288-pixel arrays. The total number of cases was 300.
We first cropped Digital Imaging and Communications in Medicine (DICOM)-formatted images with a 12 bit contrast resolution to 2022 × 2022 pixels centered on the lung region. The images were converted into tagged image file format (TIFF) images with 1024 × 1024 pixels using the bilinear interpolation method. The image pair consisting of 130 kV and 60 kV images from each patient was input to AI–DES after being normalized to a range of 0–1 for training. We used 240 pairs of images for training, 30 pairs for validation, and 30 pairs for testing. ImageJ software (1.53e, National Institutes of Health, Bethesda, MD, USA) was used to set up these datasets.

2.3. Training Environment and Parameter Settings

We used an Intel Core (TM) i7-9700K CPU and an NVIDIA GeForce RTX 2080 with 8 GB GPU memory for training. GPU acceleration was enabled using CUDA version 10.0.130, and cuDNN version 7.4.1.5-1+cuda10.0 was utilized. The implementation was performed using Python 3.7.10 and the PyTorch 1.10.0 framework on an Ubuntu 18.04.4 LTS operating system. We set the maximum number of epochs to 4000 and the batch size to 2. Adam optimization was used with the following momentum parameters: β 1 = 0.5 and β 2 = 0.999. We dynamically adjusted the learning rates as the training progressed. The learning rate of the generator started at 0.002 and decreased linearly by 0.002/4000 per epoch. The learning rate of the discriminator also decreased linearly by 0.02/4000 per epoch, starting from 0.02.

2.4. Performance Evaluation

We evaluated similarity between real and virtually generated 60 kV images for the test dataset cases. In addition, image quality of soft-tissue- and bone-enhanced images generated by our AI-DES was evaluated based on their similarity to those obtained by Discovery XR656, which is assumed to be the ground truth. Frèchet inception distance (FID) has been widely used for performance evaluation of GAN models [40,41]. However, this metric measures distances between synthetic and real data distributions; that is, FID evaluates not the similarity between each real and fake sample but the entire similarity between the two groups. Hence, we evaluated the similarity between each generated image and its corresponding ground truth using the peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [42,43], and multiscale SSIM (MS-SSIM) [44] instead of FID.
PSNR is an index based on the perceived sensitivity of noise components. It calculates the noise ratio relative to the maximum value between two images in decibels, as follows:
P S N R = 20 log 10 P m a x M S E ,
where MSE is the mean square error (MSE) between two images, and P m a x is the maximum value of the image pixels. In this study, P m a x was set to 1 because we normalized the pixel values of each image dataset, as mentioned previously in Section 2.3. The higher the value of PSNR, the more similar the two images are.
SSIM assumes image similarity using three components of brightness, contrast, and structure [42]. It has been reported that SSIM is more consistent with human perception and subjective evaluation than PSNR [45]. The images become more similar when the SSIM value is closer to 1. SSIM is calculated by dividing each region of interest (ROI) and averaging the respective SSIM values to estimate the overall similarity between the two entire images ( x , y ). We set the ROI size to 3 × 3 in this study. The calculation of SSIM between two corresponding ROIs is defined as follows:
S S I M R O I ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
where μ x and μ y are the local averages, σ x and σ y are the local standard deviations, and σ x y is local covariance. C 1 and C 2 are given by
C 1 = ( K 1 L ) 2 ,
C 2 = ( K 2 L ) 2 ,
where K 1 = 0.01, K 2 = 0.03, and L = 1 were used in this study.
MS-SSIM has been introduced as an alternative metric of SSIM to evaluate image details at various resolutions [44]. It can overcome the shortcomings of SSIM, which tends to underestimate spatial translation and overestimate image blurring [46]. MS-SSIM is computed by combining the three components of SSIM on multiple scales, as follows:
S S I M ( x , y ) = [ l M ( x , y ) ] α M · j = 1 M [ c j ( x , y ) ] β j [ s j ( x , y ) ] γ j .
The two images ( x , y ) are iteratively low-pass-filtered and downsampled by a factor of 2. The scale of the original image is 1, while that of the most reduced image is M. The brightness is compared only at scale M, and we refer to this as l M ( x , y ) . The contrast and structure components are compared at each scale, denoted as c j ( x , y ) and s j ( x , y ) , respectively, for the j th scale. Wang et al. [44] obtained five-scale parameters in which the SSIM scores agreed with subjective assessments: β 1 = γ 1 = 0.0448, β 2 = γ 2 = 0.2856, β 3 = γ 3 = 0.3001, β 4 = γ 4 = 0.2363, and α 5 = β 5 = γ 5 = 0.1333. We set M = 5 and used these five-scale parameters in the present study.
We also subjectively evaluated the image quality of the soft-tissue- and bone-enhanced images generated by AI-DES in comparison to those obtained using Discovery XR656. The subjective evaluation was conducted by three authors (A.Y., A.K., and T.I.), who are all radiological technologists, to determine whether the both images of identical patients looked similar.

3. Results

3.1. Generated Virtual Low-Energy Images

The real and 60 kV images virtually generated by our AI network for four test cases are presented in Figure 4a–d. The real and generated images look quite similar to each other. The calculated PSNR, SSIM, and MS-SSIM values for each case are shown at the bottom of the figure. The average PSNR, SSIM, and MS-SSIM values across all test cases were 33.8 dB, 0.984, and 0.957, respectively.

3.2. Soft Tissue and Bone Images

Figure 5 shows soft-tissue- and bone-enhanced images generated using Equation (10) for four test cases. The weight factors ( ω ) used in the weighted subtraction are presented in the lower-right corner of each image. As shown in the left column in Figure 5, the soft-tissue-enhanced images demonstrate that the soft tissues were well-retained, and bone tissues were effectively suppressed, although the edges of the clavicles, ribs, and spine are faintly presented. On the other hand, as seen in the right column in Figure 5, the bone-enhanced images exhibited relatively suppressed soft tissues. However, the enhanced lower thoracic and lumbar spine are poorly visualized, as they appear to be blacked-out.
Table 3 presents the average and standard deviation of PSNR, SSIM, and MS-SSIM values across all test cases, evaluating the similarity between soft-tissue-enhanced images obtained using AI-DES and Discovery XR656, as well as the similarity between bone-enhanced images. Additionally, Table 3 includes weight factors used in weighted image subtraction of AI-DES, as well as the similarity indices between the real and virtual 60 kV images, as explained in Section 3.1. The PSNR, SSIM, and MS-SSIM values for soft-tissue- and bone-enhanced images are significantly lower than the indices for 60 kV images. Table 3 also demonstrates that the values of ω varied slightly among patients, but ω values for soft tissue enhancements were consistently higher than those for bone enhancements.
Figure 6 and Figure 7 compare our soft-tissue- and bone-enhanced images, which were generated from the real 130 kV and virtual 60 kV images, to those obtained using Discovery XR656 for the respective cases. The weight factors and calculated PSNR, SSIM, and MS-SSIM values are also shown for the soft-tissue- and bone-enhanced images. While the Discovery system uses real 60 kV images in the subtraction process, our AI-DES utilizes virtually generated 60 kV images. Not only in these two cases but most test cases, the soft-tissue-enhanced images demonstrated that the bone shadows within the lung fields were successfully suppressed in both systems. However, our soft-tissue-enhanced images contained artifacts, implying the presence of thoracic and lumbar spines. Furthermore, in some cases, the mediastinum or liver area appeared overly bright in our soft-tissue-enhanced images when adjusting the contrast within the lung fields to match that in the images produced by the Discovery system, as particularly seen in the case of Figure 7.
Subsequently, as compared in Figure 6 and Figure 7, we confirmed that the bone images produced by the Discovery system exhibited remarkably more selective enhancement for bone tissues across the entire image. In contrast, our bone-enhanced images presented comparably enhanced ribs but contained artifacts in and around the region where the lower thoracic and lumbar spines should be depicted.
Figure 8 and Figure 9 also compare the soft-tissue- and bone-enhanced images generated by AI-DES to those generated by the Discovery system for other test cases, as well as enlarged views of specific areas. The enlarged views revealed that the images generated by AI-DES (the upper row in Figure 8b and Figure 9b) exhibited superior sharpness and noise characteristics, particularly in the bone-enhanced images, compared to those generated by the Discovery system (the lower row in Figure 8b and Figure 9b). Alternatively, the soft-tissue-enhanced images generated by the Discovery system showed better contrast, particularly in depicting pulmonary vessels and soft tissue lesions, and more effectively suppressed bone shadows.
Taken together, our bone suppression within the lung fields was relatively successful, although the similarity indices were not substantially high. In other words, AI-DES was able to selectively enhance soft tissues, especially within lung fields, comparable to the existing DES system using only high-energy images. Nonetheless, bone edge artifacts and excessive contrast were exhibited in the mediastinum and liver areas. The red arrows in Figure 6 and Figure 7 indicate pulmonary lesions. Although these lesions can already be easily observed in the real 130 kV images, both AI-DES and the Discovery system successfully enhanced the lesions in the soft-tissue-enhanced images.

4. Discussion

The main purpose of this study was to produce soft-tissue- and bone-enhanced images from only high-energy images by developing an AI-DES system. The performance of the AI model was assessed by image similarity between the generated and real low-energy images. The overall performance of the AI-DES system was evaluated by comparing the image quality of the soft-tissue- and bone-enhanced images with those produced by a clinically applied DES system.
As shown in Figure 4, our AI–DES successfully generated virtual low-energy images that closely resembled the real low-energy images. The image similarity indicated a high level of the indices (PSNR = 33.8 dB, SSIM = 0.984 and MS-SSIM = 0.957). Our system achieved this result without the need for a specific imaging detector with a metal plate or multiple X-ray exposures. Consequently, AI-DES has the potential to reduce the image noise and X-ray dose compared to existing DES systems. Additionally, its motion-artifact-free nature resulting from avoidance of multiple exposures makes AI-DES particularly valuable for elderly or critically ill patients who may have difficulty in holding their breath.
Furthermore, in the soft-tissue-enhanced images obtained via the weighted subtraction process, bone tissues within the lung fields were effectively suppressed, although faint residual bone edges or shadows were observed. We also confirmed that the contrast of pulmonary lesions was clearly enhanced in some cases, as indicated by red arrows in Figure 6 and Figure 7. Nevertheless, the soft tissue contrast was slightly inferior to that of the clinical system, as shown in Figure 8b and Figure 9b. Alternatively, AI-DES demonstrated advantages in terms of sharpness and noise characteristics. Overall, we subjectively verified that the image quality in the lung fields is almost comparable to that in the clinically applied system. However, these limited cases provide only weak evidence regarding the usefulness of AI-DES for improving lesion detectability. Therefore, further investigation is needed to determine whether AI-DES can truly improve the detectability for a larger number of cases.
The similarity indices (PSNR = 21.1 dB, SSIM = 0.711, and MS-SSIM = 0.794) for our soft-tissue-enhanced images were considerably lower than the results of existing AI models for bone suppression in chest radiography [23,25,28,29,30,31]. Zhou et al., generated bone-suppressed chest radiographs with a resolution of 256 × 256 using a cGAN-based model citeref-DcGANand reported an average PSNR of 35.5 dB and an SSIM of 0.975 for the similarity between the generated images and the ground truth. Rajaraman et al., also developed a CNN model called DeBoNet, which suppresses bones in chest radiographs with a resolution of 256 × 256 [30]. They achieved image similarity to the ground truth produced with commercial software, with an average PSNR of 36.8 dB, an SSIM of 0.947, and an MS-SSIM of 0.985. The low similarity indices of our soft-tissue-enhanced images can be attributed to bone edge artifacts and excessive contrast outside the lung fields, as well as the presence of black or white areas beyond the body contours where pixel values are nearly 0 or extremely high. Undoubtedly, our generated images need improvements by artifact reduction. However, a more substantially meaningful comparative analysis would have been provided by excluding the outer areas of the body contours for similarity index calculation. Additionally, it should be noted that our image generation was accomplished with a resolution of 1024 × 1024, which is certainly a greater challenge compared to 256 × 256 image synthesis.
On the other hand, the generated bone-enhanced images contained noticeable blacked-out artifacts on and around the spines, despite the successful suppression of soft tissues and the selective enhancement of ribs. The ribs and clavicles demonstrated significantly superior sharpness and noise characteristics compared to the images obtained by the Discovery system, as shown in Figure 8b and Figure 9b. However, the visibility of the spines was much lower than that of the clinically applied system, primarily due to the presence of blacked-out artifacts. Considering the importance of selectively enhanced bone images for the detection of bone fractures or tumors, addressing this issue is crucial. In addition, some of the soft-tissue-enhanced images contained faint bone shadows, as previously mentioned. We attribute these issues to the following three factors.
First, the training dataset was limited to only 240 cases. Training the AI network on a larger image dataset may address these issues, although our AI network has already produced virtual low-energy images that closely resemble real images. In future studies, we intend to investigate whether the training dataset contains misalignments between high- and low-energy images owing to patient motion. Removing misaligned data may further improve the performance of the AI network.
Secondly, we wonder whether the weight factors used in the subtraction process were not optimized. We determined the individual values of ω for each patient to generate more selectively enhanced soft tissue and bone images. However, these values may not have been optimal. Moreover, bone thickness and density vary across different body locations, even within the same patient. Accordingly, it is possibly challenging to keep all bone shadows but completely eliminate soft tissues in the entire image with a single weight factor. Even so, we believe that selectively enhanced tissue images aid in lesion detection if target tissues were effectively enhanced. Therefore, we will attempt to selectively emphasize target tissues by optimizing the weight factor in future studies.
Thirdly, the effect of quantization error was perhaps more prominently visualized through log amplification in the subtraction process. Alternatively, the cause of the artifacts may be attributed to the discrepancy between the virtual and real low-energy images, despite their high similarity indices. We noticed that the location of black artifacts around lower thoracic and lumbar spines in bone-enhanced images corresponded to areas with excessively high X-ray absorption, where the pixel values ranged from 0 to 2 in real high-energy and virtual low-energy images. We speculate that numerical errors of such small values are further enhanced by logarithmic conversion, resulting in noticeable artifacts. We performed the subtraction process using 12 bit image data, but in future studies, we plan to use images with higher contrast resolution to eliminate artifacts.
Here, we discuss the computational complexity of our AI-DES. We adopted the widely used pix2pix with a few changes, so its implementation was not a challenging task. Despite using only 240 cases for training, the similarity between the produced and target images was high enough to support stable performance. Although the network requires paired data of high- and low-energy images, the preprocessing is far from a complex computation. However, due to the high resolution of the generated images (1024 × 1024), the batch size had to be limited to a maximum of two to prevent an excessive GPU load. Next, the AI-DES requires the application of a weighted image subtraction process to the images generated by the AI network. In our current method, the weight factors are determined manually through visual inspection in order to emphasize the desired tissues to the greatest extent possible. Taken together, we consider that the construction of our AI network is no more complicated than those proposed in existing studies to directly generate bone-suppressed images [23,24,25,26,27,28,29,30,31]. However, the image subtraction process, which is the latter part of AI-DES, requires human intervention and time rather than computational complexity. While we also aim to automate the subtraction process in future work, we anticipate that it will involve complex computations, as attempted by Do et al. [19].
This study is also subject to some limitations, as described below. First, we used anonymized image data and did not include any patient information, such as gender, age, presence or absence of lesions, or medical history. Evaluating the performance separately according to various categories may provide useful insights for further improvements in this development. Particularly, a comparison of the performance between normal and diseased patients would be valuable. Next, a bias may have been introduced introduced in this study due to the use of image data collected at a single site using a specific imaging system. It is necessary to verify the performance with other datasets in future studies. Moreover, the image quality of the tissue-enhanced images generated by AI-DES was assessed subjectively by only three radiological technologists. Future studies should involve radiologists or thoracic physicians to evaluate the image quality more comprehensively.
To conclude this paper, we present one more points of superiority of the AI-DES system. It is possible to create selectively enhanced images of tissues with any linear attenuation coefficient by adjusting the weight factor in the subtraction process. Although this initial development report was focused on generating soft-tissue- and bone-enhanced images, in future work, we aim to generate enhanced images targeting specific lesions for individual patients. This approach will be feasible due to the utilization of virtual low-energy images, since it differs from existing image processing approaches that directly create bone-suppressed images [23,24,25,26,27,28,29,30,31].

5. Conclusions

Our developed AI–DES successfully generated virtual low-energy images from high-energy images obtained in routine radiography. We demonstrated that the virtual low-energy images have a high similarity to real images. Additionally, the AI-DES achieved the production of soft-tissue- and bone-enhanced images through weighted subtraction processing. The soft-tissue-enhanced images showed comparable quality, especially within lung fields, to those produced by the existing DES system while avoiding difficulties such as increased noise and exposure dose increments. The bone-enhanced images showcased advantages in terms of sharpness and noise characteristics, although noticeable artifacts on and around lower thoracic and lumbar spines need to be addressed. In future work, we aim to improve the image quality, particularly in bone-enhanced image generation, by making modifications to the AI-DES. It is also essential to evaluate the performance by involving radiologists or thoracic physicians for a wide range of image cases.

Author Contributions

Conceptualization, T.I. and T.T.; methodology, T.T. and T.I.; software, T.T. and A.Y.; validation, A.K. and A.Y.; formal analysis, A.K.; investigation, A.K., A.Y. and T.T.; resources, M.S.; data curation, M.S.; writing—original draft preparation, A.K. and A.Y.; writing—review and editing, T.I.; visualization, A.K and A.Y.; supervision, T.I.; project administration, T.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

All images used in this study were approved by the Research Ethics Review Committee of Kitasato University Hospital and Osaka University Graduate School of Medicine.

Informed Consent Statement

A waiver of informed consent was approved by the Research Ethics Review Committee of Kitasato University Hospital and Osaka University Graduate School of Medicine because of the retrospective nature of this study. According to Japan’s Ethical Guidelines for Medical and Health Research Involving Human Subjects, patients were given the opportunity to “opt out”.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, J.; Deng, Y.; Tin, M.S.; Lok, V.; Ngai, C.H.; Zhang, L.; Lucero-Prisno, D.E., 3rd; Xu, W.; Zheng, Z.J.; Elcarte, E.; et al. Distribution, risk factors, and temporal trends for lung cancer incidence and mortality: A global analysis. Chest 2022, 161, 1101–1111. [Google Scholar] [CrossRef]
  2. Panunzio, A.; Sartori, P. Lung cancer and radiological imaging. Curr. Radiopharm. 2020, 13, 238–242. [Google Scholar] [CrossRef] [PubMed]
  3. Ning, J.; Ge, T.; Jiang, M.; Jia, K.; Wang, L.; Li, W.; Chen, B.; Liu, Y.; Wang, H.; Zhao, S.; et al. Early diagnosis of lung cancer: Which is the optimal choice? Aging 2021, 13, 6214–6227. [Google Scholar] [CrossRef]
  4. Huo, J.; Shen, C.; Volk, R.J.; Shih, Y.T. Use of CT and chest radiography for lung cancer screening before and after publication of screening guidelines: Intended and unintended uptake. JAMA Intern. Med. 2017, 177, 439–441. [Google Scholar] [CrossRef] [PubMed]
  5. Adams, S.J.; Stone, E.; Baldwin, D.R.; Vliegenthart, R.; Lee, P.; Fintelmann, F.J. Lung cancer screening. Lancet 2022, 401, 390–408. [Google Scholar] [CrossRef] [PubMed]
  6. Wender, R.; Fontham, E.T.; Barrera, E., Jr.; Colditz, G.A.; Church, T.R.; Ettinger, D.S.; Etzioni, R.; Flowers, C.R.; Gazelle, G.S.; Kelsey, D.K.; et al. American Cancer Society lung cancer screening guidelines. CA Cancer J. Clin. 2013, 63, 107–117. [Google Scholar] [CrossRef] [Green Version]
  7. Wood, D.E.; Kazerooni, E.A.; Baum, S.L.; Eapen, G.A.; Ettinger, D.S.; Hou, L.; Jackman, D.M.; Klippenstein, D.; Kumar, R.; Lackner, R.P.; et al. Lung cancer screening, Version 3.2018, NCCN Clinical Practice Guidelines in Oncology. J. Natl. Compr. Cancer Netw. 2008, 98, 1602–1607. [Google Scholar] [CrossRef] [PubMed]
  8. Toyoda, Y.; Nakayama, T.; Kusunoki, Y.; Iso, H.; Suzuki, T. Sensitivity and specificity of lung cancer screening using chest low-dose computed tomography. Br. J. Cancer 2018, 16, 412–441. [Google Scholar] [CrossRef]
  9. Stitik, F.P.; Tockman, M.S. Radiographic screening in the early detection of lung cancer. Radiol. Clin. N. Am. 1978, 16, 347–366. [Google Scholar]
  10. Li, F.; Engelmann, R.; Doi, K.; MacMahon, H. Improved detection of small lung cancers with dual-energy subtraction chest radiography. AJR Am. J. Roentgenol. 2008, 190, 886–891. [Google Scholar] [CrossRef]
  11. Gomi, T.; Nakajima, M. Dual-energy subtraction X-ray digital tomosynthesis: Basic physical evaluation. Open J. Med. Imaging 2012, 2, 111–117. [Google Scholar] [CrossRef] [Green Version]
  12. MacMahon, H.; Li, F.; Engelmann, R.; Roberts, R.; Armato, S. Dual energy subtraction and temporal subtraction chest radiography. J. Thorac. Imaging 2008, 23, 77–85. [Google Scholar] [CrossRef]
  13. Kuhlman, J.E.; Collins, J.; Brooks, G.N.; Yandow, D.R.; Broderick, L.S. Dual-energy subtraction chest radiography: What to look for beyond calcified nodules. Radiographics 2006, 26, 79–92. [Google Scholar] [CrossRef]
  14. Oda, S.; Awai, K.; Funama, Y.; Utsunomiya, D.; Yanaga, Y.; Kawanaka, K.; Yamashita, Y. Effects of dual-energy subtraction chest radiography on detection of small pulmonary nodules with varying attenuation: Receiver operating characteristic analysis using a phantom study. Jpn. J. Radiol. 2010, 28, 214–219. [Google Scholar] [CrossRef]
  15. Oda, S.; Awai, K.; Funama, Y.; Utsunomiya, D.; Yanaga, Y.; Kawanaka, K.; Nakaura, T.; Hirai, T.; Murakami, R.; Nomori, H.; et al. Detection of small pulmonary nodules on chest radiographs: Efficacy of dual-energy subtraction technique using flat-panel detector chest radiography. Clin. Radiol. 2010, 65, 609–615. [Google Scholar] [CrossRef] [PubMed]
  16. Manji, F.; Wang, J.; Norman, G.; Wang, Z.; Koff, D. Comparison of dual energy subtraction chest radiography and traditional chest X-rays in the detection of pulmonary nodules. Quant. Imaging Med. Surg. 2016, 6, 1–5. [Google Scholar] [PubMed]
  17. Van der Heyden, B. The potential application of dual-energy subtraction radiography for COVID-19 pneumonia imaging. Br. J. Radiol. 2021, 94, 20201384. [Google Scholar] [CrossRef] [PubMed]
  18. Fukao, M.; Kawamoto, K.; Matsuzawa, H.; Honda, O.; Iwaki, T.; Doi, T. Optimization of dual-energy subtraction chest radiography by use of a direct-conversion flat-panel detector system. Radiol. Phys. Technol. 2015, 8, 46–52. [Google Scholar] [CrossRef] [PubMed]
  19. Do, Q.; Seo, W.; Shin, C.W. Automatic algorithm for determining bone and soft-tissue factors in dual-energy subtraction chest radiography. Biomed. Signal Process. Control 2023, 80, 104354. [Google Scholar] [CrossRef]
  20. Vock, P.; Szucs-Farkas, Z. Dual energy subtraction: Principles and clinical applications. Digit. Radiogr. 2009, 72, 231–237. [Google Scholar] [CrossRef] [PubMed]
  21. Kim, D.W.; Park, J.; Kim, J.; Kim, H.K. Noise-reduction approaches to single-shot dual-energy imaging with a multilayer detector. J. Instrum. 2019, 14, C01021. [Google Scholar] [CrossRef]
  22. Shunkov, Y.E.; Kobylkin, I.S.; Prokhorov, A.V.; Pozdnyakov, D.V.; Kasiuk, D.M.; Nechaev, V.A.; Alekseeva, O.M.; Naumova, D.I.; Dabagov, A.R. Motion artefact reduction in dual-energy radiography. Biomed. Eng. 2022, 55, 415–419. [Google Scholar] [CrossRef]
  23. Hong, G.S.; Do, K.H.; Son, A.Y.; Jo, K.W.; Kim, K.P.; Yun, J.; Lee, C.W. Value of bone suppression software in chest radiographs for improving image quality and reducing radiation dose. Eur. Radiol. 2021, 31, 5160–5171. [Google Scholar] [CrossRef] [PubMed]
  24. Matsubara, N.; Teramoto, A.; Saito, K.; Fujita, H.; Naumova, D.I. Bone suppression for chest X-ray image using a convolutional neural filter. Phys. Eng. Sci. Med. 2020, 43, 97–108. [Google Scholar] [CrossRef]
  25. Liu, Y.; Liu, M.; Xi, Y.; Qin, G.; Shen, D.; Yang, W. Generating dual-energy subtraction soft-tissue images from chest radiographs via bone edge-guided GAN. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; pp. 678–687. [Google Scholar]
  26. Bae, K.; Oh, D.Y.; Yun, I.D.; Jeon, K.N. Bone suppression on chest radiographs for pulmonary nodule detection: Comparison between a generative adversarial network and dual-energy subtraction. Korean J. Radiol. 2022, 23, 139–149. [Google Scholar] [CrossRef]
  27. Cho, K.; Seo, J.; Kyung, S.; Kim, M.; Hong, G.-S.; Kim, N. Bone suppression on pediatric chest radiographs via a deep learning-based cascade model. Comput. Methods Programs Biomed. 2022, 215, 106627. [Google Scholar] [CrossRef]
  28. Zhou, Z.; Zhou, L.; Shen, K. Dilated conditional GAN for bone suppression in chest radiographs with enforced semantic features. Med. Phys. 2020, 47, 6207–6215. [Google Scholar] [CrossRef]
  29. Zarshenas, A.; Liu, J.; Forti, P.; Suzuki, K. Separation of bones from soft tissue in chest radiographs: Anatomy-specific orientation-frequency-specific deep neural network convolution. Med. Phys. 2019, 46, 2232–2242. [Google Scholar] [CrossRef]
  30. Rajaraman, S.; Cohen, G.; Spear, L.; Folio, L.; Antani, S. DeBoNet: A deep bone suppression model ensemble to improve disease detection in chest radiographs. PLoS ONE 2022, 17, e0265691. [Google Scholar] [CrossRef]
  31. Rani, G.; Misra, A.; Dhaka, V.S.; Zumpano, E.; Vocaturo, E. Spatial feature and resolution maximization GAN for bone suppression in chest radiographs. Comput. Methods Programs Biomed. 2022, 224, 107024. [Google Scholar] [CrossRef]
  32. Isora, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
  33. Yoshida, N.; Kageyama, H.; Akai, H.; Yasaka, K.; Sugawara, H.; Okada, Y.; Kunimatsu, A. Motion correction in MR image for analysis of VSRAD using generative adversarial network. PLoS ONE 2022, 17, e0274576. [Google Scholar] [CrossRef]
  34. Sun, J.; Du, Y.; Li, C.; Wu, T.-H.; Yang, B.; Mok, G.S.P. Pix2Pix generative adversarial network for low dose myocardial perfusion SPECT denoising. Quant. Imaging Med. Surg. 2022, 12, 3539–3555. [Google Scholar] [CrossRef]
  35. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  36. junyanz/pytorch-CycleGAN-and-pix2pix. Available online: https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix (accessed on 5 June 2023).
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  38. Pathak, D.; Krähenbühl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context encoders: Feature learning by inpainting. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2536–2544. [Google Scholar]
  39. Hayashi, N.; Taniguchi, A.; Noto, K.; Shimosegawa, M.; Ogura, T.; Doi, K. Development of a digital chest phantom for studies on energy subtraction techniques. Nihon Hoshasen Gijutsu Gakkai Zasshi 2014, 70, 191–198. (In Japanese) [Google Scholar] [CrossRef] [Green Version]
  40. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6626–6637. [Google Scholar]
  41. Borji, A. Pros and cons of GAN evaluation measures. arXiv 2018, arXiv:1802.03446. [Google Scholar] [CrossRef] [Green Version]
  42. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error measurement to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Hóre, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  44. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the 37th IEEE Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  45. Umme, S.; Morium, A.; Mohammad, S.U. Image quality assessment through FSIM, SSIM, MSE and PSNR—A comparative study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar]
  46. Mudeng, V.; Kim, M.; Choe, S. Prospects of structual similarity index for medical image analysis. Appl. Sci. 2022, 12, 3754. [Google Scholar] [CrossRef]
Figure 1. AI network diagram using pix2pix. The generator attempts to produce virtual low-energy images that resemble the real low-energy images from high-energy images. The discriminator aims to distinguish between the pairs containing virtual images and the real image pairs.
Figure 1. AI network diagram using pix2pix. The generator attempts to produce virtual low-energy images that resemble the real low-energy images from high-energy images. The discriminator aims to distinguish between the pairs containing virtual images and the real image pairs.
Applsci 13 07220 g001
Figure 2. Generator network used in AI-DES. The network has a 16-layer U-Net structure. The skip connections concatenate the mirrored encoder and decoder layers to recover high-frequency components in the generated image quality.
Figure 2. Generator network used in AI-DES. The network has a 16-layer U-Net structure. The skip connections concatenate the mirrored encoder and decoder layers to recover high-frequency components in the generated image quality.
Applsci 13 07220 g002
Figure 3. Discriminator network used in AI-DES. The network comprises five convolutional neural network layers. It takes either a pair of real high-energy and virtual low-energy images or a pair of real high-energy and low-energy images as input. The network then outputs a feature map that determines whether the pair is real or fake.
Figure 3. Discriminator network used in AI-DES. The network comprises five convolutional neural network layers. It takes either a pair of real high-energy and virtual low-energy images or a pair of real high-energy and low-energy images as input. The network then outputs a feature map that determines whether the pair is real or fake.
Applsci 13 07220 g003
Figure 4. Examples of low-energy images virtually generated by our trained AI network. Four test cases (ad) are presented here. The similarity indices between real and virtual images are presented at the bottom of each figure.
Figure 4. Examples of low-energy images virtually generated by our trained AI network. Four test cases (ad) are presented here. The similarity indices between real and virtual images are presented at the bottom of each figure.
Applsci 13 07220 g004
Figure 5. Examples of soft-tissue- and bone-enhanced images generated by AI-DES for four test cases (ad). The weight factors used in the subtraction process are presented in the lower-right corner of each image.
Figure 5. Examples of soft-tissue- and bone-enhanced images generated by AI-DES for four test cases (ad). The weight factors used in the subtraction process are presented in the lower-right corner of each image.
Applsci 13 07220 g005
Figure 6. An example of soft-tissue- and bone-enhanced images generated by AI-DES with real 130 kV and virtual 60 kV images in comparison to the enhanced images generated by the Discovery XR656 system. The weight factors and similarity indices between the enhanced images generated by AI-DES and Discovery XR656 are presented. The red arrow indicates a pulmonary lesion.
Figure 6. An example of soft-tissue- and bone-enhanced images generated by AI-DES with real 130 kV and virtual 60 kV images in comparison to the enhanced images generated by the Discovery XR656 system. The weight factors and similarity indices between the enhanced images generated by AI-DES and Discovery XR656 are presented. The red arrow indicates a pulmonary lesion.
Applsci 13 07220 g006
Figure 7. Another example of soft-tissue- and bone-enhanced images generated by AI-DES with real 130 kV and virtual 60 kV images in comparison to the enhanced images generated by the Discovery XR656 system. The weight factors and similarity indices between the enhanced images generated by AI-DES and Discovery XR656 are presented. The red arrows indicate a pulmonary lesion.
Figure 7. Another example of soft-tissue- and bone-enhanced images generated by AI-DES with real 130 kV and virtual 60 kV images in comparison to the enhanced images generated by the Discovery XR656 system. The weight factors and similarity indices between the enhanced images generated by AI-DES and Discovery XR656 are presented. The red arrows indicate a pulmonary lesion.
Applsci 13 07220 g007
Figure 8. Comparison of soft-tissue- and bone-enhanced images obtained by AI-DES and Discovery XR656 for a test case. (a) Overall views of the enhanced images. The weight factors and similarity indices between the enhanced images generated the two systems are also presented. (b) Enlarged views of the specific areas are enclosed by the orange and blue dotted boxes in the overall views.
Figure 8. Comparison of soft-tissue- and bone-enhanced images obtained by AI-DES and Discovery XR656 for a test case. (a) Overall views of the enhanced images. The weight factors and similarity indices between the enhanced images generated the two systems are also presented. (b) Enlarged views of the specific areas are enclosed by the orange and blue dotted boxes in the overall views.
Applsci 13 07220 g008
Figure 9. Comparison of soft-tissue- and bone-enhanced images obtained by AI-DES and Discovery XR656 for another test case. (a) Overall views of the enhanced images. (b) Enlarged views of specific areas are enclosed by orange and blue dotted boxes in the overall views.
Figure 9. Comparison of soft-tissue- and bone-enhanced images obtained by AI-DES and Discovery XR656 for another test case. (a) Overall views of the enhanced images. (b) Enlarged views of specific areas are enclosed by orange and blue dotted boxes in the overall views.
Applsci 13 07220 g009
Table 1. Generator architecture in AI-DES.
Table 1. Generator architecture in AI-DES.
TypeNorm 1, DropoutActivationInput Shape 2Output Shape 2
EncoderLayer1Conv2d (4,2,1)LekyReLU1024 × 1024 × 3512 × 512 × 64
Layer2BN512 × 512 × 64256 × 256 × 128
Layer3256 × 256 × 128128 × 128 × 256
Layer4128 × 128 × 25664 × 64 × 512
Layer564 × 64 × 51232 × 32 × 512
Layer632 × 32 × 51216 × 16 × 512
Layer716 × 16 × 5128 × 8 × 512
Layer88 × 8 × 5124 × 4 × 512
DecoderLayer9Deconv2d (4,2,1)BN4 × 4 × 5128 × 8 × 512
Layer10ReLU+Deconv2d (4,2,1)BN+Dropout8 × 8 × 51216 × 16 × 512
Layer1116 × 16 × 51232 × 32 × 512
Layer123 2 × 32 × 51264 × 64 × 512
Layer13BN64 × 64 × 512128 × 128 × 256
Layer14128 × 128 × 256256 × 256 × 128
Layer15256 × 256 × 128512 × 512 × 64
Layer16Tanh512 × 512 × 641024 × 1024 × 3
1 Normalization. 2 Width × height × channel.
Table 2. Discriminator architecture in AI-DES.
Table 2. Discriminator architecture in AI-DES.
TypeNormalizationActivationInput Shape 1Output Shape 1
Layer1Conv2d (4,2,1)LekyReLU1024 × 1024 × 6512 × 512 × 64
Layer2BN512 × 512 × 64256 × 256 × 128
Layer3256 × 256 × 128128 × 128 × 256
Layer4Conv2d (4,1,1)128 × 128 × 256127 × 127 × 512
Layer5127 × 127 × 512126 × 126 × 1
1 Width × height × channel.
Table 3. Similarity indices (average ± standard deviation) for all test cases and weight factors used to produce soft-tissue- and bone-enhanced images in AI-DES.
Table 3. Similarity indices (average ± standard deviation) for all test cases and weight factors used to produce soft-tissue- and bone-enhanced images in AI-DES.
PSNRSSIMMS-SSIMWeight Factor
60 kV images (virtual and real)33.8 ± 5.390.984 ± 0.005540.957 ± 0.0514
Soft tissue images (AI-DES and Discovery)21.1 ± 2.560.711 ± 0.05510.794 ± 0.06402.47 ± 0.159
Bone images (AI-DES and Discovery)18.3 ± 1.970.433 ± 0.08270.571 ± 0.1011.52 ± 0.102
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yamazaki, A.; Koshida, A.; Tanaka, T.; Seki, M.; Ishida, T. Development of Artificial Intelligence-Based Dual-Energy Subtraction for Chest Radiography. Appl. Sci. 2023, 13, 7220. https://doi.org/10.3390/app13127220

AMA Style

Yamazaki A, Koshida A, Tanaka T, Seki M, Ishida T. Development of Artificial Intelligence-Based Dual-Energy Subtraction for Chest Radiography. Applied Sciences. 2023; 13(12):7220. https://doi.org/10.3390/app13127220

Chicago/Turabian Style

Yamazaki, Asumi, Akane Koshida, Toshimitsu Tanaka, Masashi Seki, and Takayuki Ishida. 2023. "Development of Artificial Intelligence-Based Dual-Energy Subtraction for Chest Radiography" Applied Sciences 13, no. 12: 7220. https://doi.org/10.3390/app13127220

APA Style

Yamazaki, A., Koshida, A., Tanaka, T., Seki, M., & Ishida, T. (2023). Development of Artificial Intelligence-Based Dual-Energy Subtraction for Chest Radiography. Applied Sciences, 13(12), 7220. https://doi.org/10.3390/app13127220

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop