Next Article in Journal
Interval Type-II Fuzzy Fault-Tolerant Control for Constrained Uncertain 2-DOF Robotic Multi-Agent Systems with Active Fault Detection
Next Article in Special Issue
Fast Semantic Segmentation of Remote Sensing Images Using a Network That Integrates Global and Local Information
Previous Article in Journal
Application of the Nicoya OpenSPR to Studies of Biomolecular Binding: A Review of the Literature from 2016 to 2022
Previous Article in Special Issue
Game-o-Meta: Trusted Federated Learning Scheme for P2P Gaming Metaverse beyond 5G Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Branch Attention Learning for Bone Age Assessment with Ambiguous Label

School of Automation (School of Artificial Intelligence), Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4834; https://doi.org/10.3390/s23104834
Submission received: 22 March 2023 / Revised: 15 May 2023 / Accepted: 15 May 2023 / Published: 17 May 2023

Abstract

:
Bone age assessment (BAA) is a typical clinical technique for diagnosing endocrine and metabolic diseases in children’s development. Existing deep learning-based automatic BAA models are trained on the Radiological Society of North America dataset (RSNA) from Western populations. However, due to the difference in developmental process and BAA standards between Eastern and Western children, these models cannot be applied to bone age prediction in Eastern populations. To address this issue, this paper collects a bone age dataset based on the East Asian populations for model training. Nevertheless, it is laborious and difficult to obtain enough X-ray images with accurate labels. In this paper, we employ ambiguous labels from radiology reports and transform them into Gaussian distribution labels of different amplitudes. Furthermore, we propose multi-branch attention learning with ambiguous labels network (MAAL-Net). MAAL-Net consists of a hand object location module and an attention part extraction module to discover the informative regions of interest (ROIs) based only on image-level labels. Extensive experiments on both the RSNA dataset and the China Bone Age (CNBA) dataset demonstrate that our method achieves competitive results with the state-of-the-arts, and performs on par with experienced physicians in children’s BAA tasks.

1. Introduction

Human growth and development have an objective judgment standard, i.e., bone age, except for physical age. The early diagnosis of growth and development anomalies in teenagers based on bone age can not only estimate growth but also detect potential endocrine diseases, thus allowing intervention measures to be adopted timely [1,2]. Bone age assessment (BAA) is usually implemented by radiologists by examining left-hand X-ray images. Currently, the most common manual methods are Greulich–Pyle (GP) atlas and Tanner Whitehouse 2 (TW2) scoring method. For the GP atlas method [3], doctors determine bone age by searching the standard reference image set for the most similar image to the X-ray image of the patients’ hand. The TW2 method [4] recognizes the specific regions of interest (ROIs) of wrist and finger bones, then scores each ROI according to development degree, and finally combines the scores to calculate bone age. However, both the GP atlas and TW2 scoring methods suffer from some limitations in BAA. First, these two methods have strong subjectivity and are easily interfered by doctors’ subjective judgment. Moreover, they involve high requirements on professional levels of doctors, and bone age can be determined accurately only if provided with long-term training and experience. Lastly, both methods are time- and labor-consuming. Currently, the GP atlas is the preferred method because the TW2 method requires determining the grade of each RoI, which increases the workload of physicians.
Deep learning (DL) has been used with tremendous success in a variety of medical image analysis tasks, such as pulmonary nodule detection [5], chest X-ray diagnosis [6], and intracranial hemorrhage detection [7]. Accordingly, scholars have proposed a series of DL techniques for BAA using the open BAA dataset Radiological Society of North America (RSNA) and achieved excellent performance. DL-based BAA methods can generally be classified into two categories: single-stage and multi-stage structures. The former one pretrains the networks such as VGG [8], ResNet [9], or Inception [10] on ImageNet [11] and then predicts the bone age directory by adding a regression layer at the end of the network. The latter first segment the hand object from original images by employing an ROI-based method with additional processing, then detects and extracts bone part regions with human prior knowledge, and finally predicts bone age through a regression layer. In comparison with the single-stage structure, the multi-stage structure achieves better performance. Although previous related research have confirmed the feasibility of deep learning in BAA with outstanding performance, they mainly train, verify, and test their network on RSNA or private dataset. None of these studies have evaluated the accuracy and generalization of these models by using external datasets. As a result, the generalization of the automatic BAA model on the target dataset declines during test, which is mainly attributed to covariate shift (P(x) varies) [12] and prior probability shift (P(y) varies) [13]. For P(x) varies, the image distributions of the source and target data vary because of the use of different X-ray machines, image transmission protocols, and image shooting modes. The variations of P(y) are mainly caused by the different growth and development stages of hand bone characteristics of patients from different races, and the “Ground truth” of images is labeled by different standards. For instance, Westerners and Easterners adopt inconsistent GP atlas, and the TW method adopts different standards for the grade labeling of each ROI for different races. Therefore, the BAA model requires a re-collected target dataset and fine-tuning based on transfer learning to improve its generalization among different races. However, this suffers from the following challenges: (1) It is hard to acquire enough accurately labeled images since many professional radiologists and material resources are needed. (2) In the radiological report, bone age is generally expressed by uncertain description or an interval, such as “approximately 12 years old”, or “aged between 13 and 14”. In comparison with the accurate labeling to months in the RSNA dataset, only ambiguous bone age labels can be obtained. (3) The multi-stage training method needs to annotate the ROIs of hand bones, resulting in extra workload. To address the above challenges, we propose an automatic BAA method based on multi-branch attention learning with ambiguous labels (MAAL-Net). Specifically, this method extracts ambiguous labels of bone age from the radiological report by using natural language processing (NLP) techniques, acquires ROIs of hand bone through the multi-branch attention learning mechanism, and predicts bone age through end-to-end training.
The major contributions of this study are as follows: (1) A novel MAAL-Net is proposed, which automatically extracts hand object and bone part regions with only image-level bone age annotation. Our method is very similar to clinical assessment techniques, which makes it transparent for clinicians. (2) Considering the high difficulties in collecting enough accurately labeled bone age data, ambiguous labels are extracted from the radiological report by using NLP techniques. These ambiguous labels are used to train the model for automatic and accurate bone age prediction. (3) To the best of our knowledge, this study is the first to train, verify, and test the automatic BAA method on datasets of different races, and it will promote the applications of automatic BAA systems worldwide.
The rest of this paper is organized as follows. The related work is reviewed in Section 2. Section 3 illustrates the details of MAAL-Net. Section 4 presents the experiments and shows the experimental results of BAA task. The conclusion is given in Section 5.

2. Related Work

Automatic BAA methods consist of three main categories. The first category is the traditional machine learning method. Some traditional methods use hand-crafted features, e.g., gray-level co-occurrence matrix (GCLM) features, local binary pattern (LBP) features, and scale-invariant feature transform (SIFT) features, to train support vector machine (SVM) or its deformation methods [14,15,16]. Besides, BoneXpert [17] is known as the most popular automatic BAA software in Europe. It carries out automatic BAA by shape-driven active appearance model and predicts bone age features extracted by the principal component analysis method. The rejection rate of poor-quality images is approximately 1%, while the accuracy of this system is around 7.8 months. In general, traditional machine learning methods employ hand-crafted features that are sensitive to image quality and have poor accuracy.
The second category adopts the ROI-free method with single-stage structure to predict bone age, which uses the whole image as the input. For example, Larson et al. [18] used ResNet50 as the backbone network and 14,036 X-ray images of hands as the training set. The network determines the bone age according to the probability for every X-ray image belonging to which month. The mean absolute error (MAE) of this network on the test set was 6.24 months. Spampinato et al. [19] proposed and tested several DL-based methods for automatic BAA. Results demonstrated an average variation between automated and manual assessments of approximately 0.8 years. Recently, Cristina et al. [20] proposed a specific identity labels-based bone age assessment (SIMBA) model, which involved the gender as well as the age of patients. Different from the other bone age assessment methods, this model calculates the relative bone age according to the difference between practical age and bone age. The SIMBA proposes a new paradigm for BAA and achieves an MAE of 5.47 months on the private dataset RHPE. However, the application of this model is limited by the facts that a patient’s age is private and medical institutions usually do not provide age data to the public. All these methods predict bone age by training end-to-end deep neural networks using the entire image as input, which results in limited accuracy and poor interpretability.
The third category adopts a ROI-based method with the multi-stage structure to predict bone age, which extracts local information and ultimately predicts bone age by introducing prior knowledge for image segmentation or hand bone ROIs detection. For instance, Escobar et al. [21] proposed the Bonet which introduced a hand detection module and pose estimation module to the BBA system. This method had an MAE of 4.14 months on the RSNA dataset by virtue of its fine granularity and precisely localized annotation. Wang et al. [22] estimated bone age by extracting the local regions of phalanges and ulna from hand images based on RCNN. However, this model only evaluated the performance of local regions. Wang et al. [23] proposed the anatomical local-aware network (ALA-NET), which learned the architecture of hands and extracted local features following the spirit TW2 method. They estimated bone age by analyzing the features of ROIs, and this model obtained an MAE of 4.49 months without ROIs annotations. Chandran et al. [24] proposed to use the segmented ROI as input to a Generative Adversarial Network (GAN) for bone age prediction. However, GAN gradient descent sometimes fails to reach Nash equilibrium, resulting in unstable training, vanishing gradients, mode collapse, and other issues. Although GANs can achieve excellent performance, this method still requires a dataset with accurately annotated ROIs. Additionally, the model is trained on a bone age dataset of Western children, which is very different from Eastern children in terms of growth and development patterns and bone age assessment standards. Therefore, this trained model cannot be directly used for bone age assessment in Eastern children. Although the multi-stage structure achieves better performance, these models have high complexity and long computational time due to the need for prior knowledge and extra annotations.
In contrast to the above studies, the proposed MAAL-Net overcomes the difficulty in acquiring exact labels by using ambiguous labels. Moreover, it enables an end-to-end neural network to detect the hand object and the most informative bone part regions without using bounding box or part annotations.

3. Methodology

As shown in Figure 1, the proposed MAAL-Net has three branches in the training stage: the raw image branch (Branch 1), the hand image branch (Branch 2), and the bone part images branch (Branch 3). Through the collaborative learning of these three branches, the network achieves a good classification ability for different scales and parts of objects as well. In the test phase with repeat experiments, Branch 2 achieves the best performance. Accordingly, we use Branch 2 as the final prediction output. Two modules are included among the three branches. One is the hand object location module (HOLM) to predict the position of the hand object, and the other is the attention part extraction module (APEM) to extract the ROIs with the most information of the hand object without using bounding-box annotations. Specifically, HOLM uses the whole image as input and generates the hand object that covers the most discriminative regions. This image is then used as the input of APEM to further crop bone part regions for training.

3.1. Construction of Ambiguous Label

Existing BAA methods adopt strong supervision hypotheses during learning and modeling: sufficient labeled samples are available during training, and the label of each sample is unique and explicit. Nevertheless, precise month-to-month bone age labels cannot be obtained in practical clinical diagnosis reports due to the inconsistent diagnostic criteria and professional levels of expertise. Common labels are ambiguous labels, such as bone age of approximately 12 years old or 12–13 years old. In other words, physicians can exclude impossible bone age, but they cannot provide an accurate and explicit bone age from several similar values. Hence, such ambiguous labels should be transformed into weak labels for learning. In the present study, ambiguous labels are transformed into label distribution.
Label distribution learning [25]: the real label distribution ( y ) is usually unavailable for an ambiguous label, which should be generated under an appropriate hypothesis. The label distribution y = ( y 1 , y 2 , . . . y C ) must satisfy the following two conditions: (1) y shall be the probability distribution, y i [ 0,1 ] and i = 1 C y i = 1 where c is the number of the classes. (2) The probability value ( y i ) related to images varies among all possible labels. Relatively certain categories shall be vested with higher probability values, while those uncertain categories shall be given with relatively lower probability values. For BAA, the hypothesized probability shall be concentrated around the bone age labeled by physicians. Hence, ambiguous labels are quantified by Gaussian distribution. For bone age, a target label set can be found, where 0 < Bone-age ≤ 240. It is quantified into an integral ordered label set L = [l1, l2,…, lc], where li is the possible prediction for y . y = { y 1 , y 2 , y 3 …, y C }, where y i is the probability that y = li. Considering that equal step length is used in the quantization of y , the distribution y is generated by the following Gaussian probability density function:
y i = p ( l i | μ , σ ) C p ( l c | μ , σ )
where p ( l i | μ , σ ) = 1 2 π σ exp ( ( l i μ ) 2 2 σ 2 ) , p is the probability density function and σ is the standard deviation. We analyze the hyper-parameter σ in Section 4.4, and set σ = 1, 2, 3, 4 respectively in this study. The label distributions of images with bone age labeled at approximately 9 years old (~108 months) and labeled at 12–13 years old (~150 months) are shown in Figure 2.

3.2. Hand Object Location Module (HOLM)

Given an input image, we first use a feature extractor to obtain representative features F R H × W × K , where K refers to the channel dimension, and H × W refers to the feature resolution. F is then used as the input of a 1 × 1 convolutional layer to obtain the feature map A R H × W × C . Specifically, the c-th ( c C ) channel of A is expressed as
A c i , j = k = 0 K - 1 W k c F i j k
where F i j k , i R H , j R W , k R K presents each pixel value of the channel k in F . W k c presents the weight of the k-th input neuron and the c-th output neurons in the 1 × 1 convolutional layer. Afterward, A is fed into a pooling layer [26]. In this study, we use the global average pooling to obtain the output y ^ R C ,
y ^ c = 1 H × W i j k = 0 K - 1 W k c F i j k
where c R C is the label category, and y ^ c denotes the probability of category c. Obviously, y ^ c is the global pooling result of A c . That is,
y ^ c = 1 H × W i j A c i , j
In a word, the activated output A of the final convolutional layer ( 1 × 1 convolutional layer) for the input image labeled c could be obtained through HOLM. Specifically, A c of the c-th channel can reflect the impact of each pixel on the ultimate prediction results. When global average pooling is applied, the classification results are only sensitive to the peak value of A c . Under such circumstances, the region where peak value of A c is located is regarded as the most discriminative region, which is called ROI.
Considering that labels are ambiguous labels in Gaussian distribution, and the probability sum of all categories is 1, the output ( y ^ ) needs to be transformed into the Gaussian distribution p by using the Softmax activation function. The p-value is formulated as
p c = exp ( y ^ c ) c = 0 C - 1 exp ( y ^ c ) , c = 0,1 , . . . , C - 1
where p c is positively related to y ^ c . Consequently, the effect of the peak values of A c on y ^ c are positively related to their effect on p c .
P c 1 H × W i j A c i , j
Finally, after the normalization of A c , we design a binary mask M to pinpoint the most discriminant areas:
M c i , j = 1 A c ( i , j ) τ 0 A c i , j < τ
where τ is the threshold size of RoI. When τ decreases, RoI becomes relatively large. Otherwise, the RoI becomes relatively smaller. Hence, a suitable RoI can be obtained by adjusting τ . In this study, we empirically set τ = 0.6.

3.3. Attention Part Extraction Module (APEM)

Since BAA is fine-grained, the accuracy is theoretically improved if bone part regions could be detected for training. In APEM, the region proposal network (RPN) [27] is used to generate the RoIs bounds ( R i ) and the confidence score ( S R i ). R i was arranged in descending order according to S R i and the top N regions { R 1 , R 2 , , R N } are acquired using the non-maximum suppression (NMS) method [28].
Specifically, the output of HOLM serves as the input of APEM. For the activation output A of HOLM, we first unify its size to (576, 576) and then feed it into the RPN to produce a series of rectangle RoIs { R 1 , R 2 , R 3 , . . , R M } . The anchors of PRN are designed with ratios of {2:3, 1:1, 3:2} and scales of {128, 160, 224, 288, 384}. In other words, each point in A generates 3 × 5 groups of R i . The normalization function is defined as
z x = x - M i n ( x ) M a x x - M i n ( x )
For the category c, we can obtain S by doing an exponential transformation on the feature map A c , denoted as
S i , j = z exp η z A c i , j , η > 0
The hyper-parameter η controls the stretching degree of A c . A higher η value indicates greater differences between each local peak and surrounding region. Under such circumstances, the mean of S under the coverage of each R i is used as the corresponding confidence score ( S R i ).
S R * = 1 n i , j S ( i , j )
In comparison with A c , S after stretching can effectively assure that the confidence core of R i at peak is higher than others, and R i will not be eliminated during NMS, thus improving the reliability of location performances of APEM. Finally, R i is arranged in descending order according to confidence score:
S R 1 S R 2 S R 3 S R M
The top N terms are used as outputs, which are { R 1 , R 2 , R 3 , . . , R N } and { S R 1 , S R 2 , S R 3 , , S RN } . Finally, the crop regions are unified to (288, 288) by bilinear interpolation.

3.4. Loss Function

Cross-entropy loss is widely used in convolutional neural network-based classification tasks. However, this work transforms ambiguous labels into label distribution for learning, resulting in the inability to apply cross-entropy loss for BAA learning directly. To train the network while making the ambiguous labels work effectively, we propose the loss functions L raw and L hand for the raw and hand image branch, respectively.
L raw = L KL ( Y , Y ^ r a w ) + L m a e ( Y , Y ^ r a w ) + T ( t ) . L t ( Y ¯ r a w , Y ^ r a w )
L hand = L KL ( Y , Y ^ h a n d ) + L m a e ( Y , Y ^ h a n d ) + T ( t ) . L t ( Y ¯ h a n d , Y ^ h a n d )
where Y ^ raw and Y ^ hand are the output results of the last Softmax layer, whose distributions are consistent with the definition of Y. In other words, y ^ i [ 0 , 1 ] and i = 1 m y ^ i = 1 . Y ¯ raw and Y ¯ hand are the values produced by assembled prediction, which are updated in each epoch. T(t) is a time-dependent weighting function. For the raw and hand image branch, Equation (12) is used as the loss function. The bone part image branch does not use temporal ensembling as the third term of loss function because of the great difference in the prediction values of each RoI. In other words, L bonePart = L KL ( Y , Y ^ bonePart ) + L m a e ( Y , Y ^ b o b e P a r t ) . Specifically, the first term of loss function aims to calculate the relevance between the predicted bone age distribution ( Y ^ ) and the ground-truth bone age distribution (Y). Different methods can be used to estimate the relevance between two distributions, e.g., Kullback–Leibler (KL) divergence is adopted in this study, as Equation (14) defines:
L KL ( Y , Y ^ ) = 1 m i = 1 m y i T log ( y i y ^ i )
The second term of the loss function aims to calculate the MAE loss between Y ^ and Y, defined as Equation (15):
L mae ( Y , Y ^ ) = 1 m i = 1 m | y i y ^ i |
The third term of the loss function is the temporal ensembling term. It makes Y ^ at present training epoch consistent with the network output results ( Y ¯ ) in different epochs. Based on the temporal ensembling [29] used for semi-supervised learning, we assemble model predictions in different epochs as auxiliary supervision information for the next epoch. In comparison with the current predictions, the training targets generated by the assembled prediction might be closer to the ground-truth label distribution. Therefore, they can be used to guide network training. The temporal ensembling loss function is defined as Equation (16):
L t ( Y ¯ , Y ^ ) = 1 m i = 1 m y ¯ i T log ( y ¯ i y ^ i )
where Y ¯ i is the assembled predictions generated in each epoch. We use the exponential moving average (EMA) of predictions in every epoch to generate the assembled predictions. The assembled predictions S and the training target Y ¯ are updated as follows:
S t = γ S t 1 + ( 1 γ ) Y ^ t
Y ¯ t + 1 = S t / ( 1 γ t )
where S t and S t - 1 denote the assembled predictions at t-th and (t − 1)-th epochs, respectively. Coefficient γ is the momentum term. As expressed in Equations (17) and (18), EMA assigns smaller weights to early prediction and increases the weights as the training epoch increases. Considering that prediction values are not present before the first epoch, Y ¯ and S are initialized as 0.
The loss function has the term T(t), which is a time-dependent weighting function and increases with the number of epochs. The prediction results are unreliable at the beginning phase of training because the network has not been trained well. As a result, the assembled prediction can only provide very limited supervision information. As the training goes on, the prediction ability of the network is strengthened gradually, and the assembled prediction can guide the network to obtain better and more accurate results. T(t) is defined as follows:
T ( t ) = T max · exp ( 5 ( 1 t ) 2 )
where Tmax is the maximum value that T(t) can reach. With the increase of time t, T(t) increases from 0 to Tmax gradually.
Ultimately, the total loss function is expressed as:
L t o t a l = L r a w + L h a n d + L b o n e P a r t s

4. Experiments and Results

4.1. Datasets and Evaluation Metric

The RSNA dataset [30] was proposed by Radiological of North America in the machine learning challenges of pediatric bone age. The RSNA dataset is composed of 12,043 X-ray images of children’s left hands which are interpreted by pediatric radiologists and recorded in the radiology report for bone age [3]. Among them, 9634 images are utilized for training, 1204 images are used for validation, and the final 1205 images are used for testing. The average patient age in the training and validation set is 127 months, while that in the test set is 132 months.
X-ray images in the CNBA dataset are obtained from Chinese children. Unlike the other BAA datasets mainly composed of Westerners, East Asians dominate the CNBA dataset. The CNBA dataset has 46% male and 54% female X-ray images, with patient ages ranging from 0 to 240 months. Ambiguous labels are extracted by NLP techniques from clinical radiological reports, including bone age and gender. The CNBA dataset contains 4526 images, of which 3620 images are utilized for training, 452 images for validation, and 453 images for testing.
The MAE between predicted bone age y ^ i and its ground truth y i is used as the performance evaluation metric, defined as
MAE = 1 N i = 1 N | y i y ^ i |
Another evaluation method follows [31] to calculate the accuracy within 6 months, 12 months, and 24 months.

4.2. Implementation Details

The implementation of MAAL-Net is very flexible. In the present study, we implement the backbone of MAAL-Net using ResNet50, which can also be replaced by other networks, such as VGG, ResNet, and Inception. In the experiment, the input size is uniformly 576 × 576, and the batch size is 32. Comparative experiments are conducted with σ ranging from 1 to 4. Finally, Branch 2 achieves the best performance when σ = 3. MAAL-Net is implemented by Pytorch and is trained on a CentOS workstation with four NVIDIA GTX 1080Ti 12G GPU. Momentum SGD is chosen as the optimizer with an initial learning rate of 0.001 and declined by a factor of 10 per 30 epochs. Finally, the model that performed best on the validation set is chosen for testing.

4.3. Experimental Results

In this section, we compare our MAAL-Net with the state-of-the-art ones. The results are listed in Table 1. Generally, the RoI-based methods [21,32,33] that detect the local regions as RoIs by annotating additional bounding boxes or key points achieve better performance. By contrast, the RoI-free methods [18,19,34,35] that use entire images as input are inferior because they only use global features. However, these RoI-free methods can save a lot of manpower and materials for annotations. The method in [36] predicted bone age by using a two-stage network, which does not require ROI annotation yet achieves comparable performance as the ROI-based method. However, this method suffers some limitations because it cannot apply end-to-end training. The proposed MAAL-Net uses the multi-branch attention learning network without bounding boxes or part annotations. In comparison with RoI-free methods, the presence of multi-branch attention gives MAAL-Net a significant improvement in prediction accuracy. In comparison with RoI-based methods, the MAAL-net automatically determines the most informative regions without bounding boxes or part annotations. In comparison with the previous method [36], the MAAL-Net allows end-to-end training in a short inference time without extra processing. The experimental results indicate that our method yields state-of-the-art performance with MAE = 4.09 on the RSNA dataset.
To our knowledge, MAAL-Net is the first to train a network with ambiguous labels, which extracts ambiguous labels directly from radiological reports as training and test sets. Considering that the bone images in the CNBA test set are labeled by an ambiguous value or interval, the accuracy of MAAL-Net is evaluated based on the deviation between the predicted and the ground-truth bone age within ±6 months, ±12 months, and ±24 months. The results are shown in Table 2. Thanks to the label distribution learning that prevents over-fitting, the accuracy of the method is significantly improved by 61% within ±6 months, 7.8% within ±12 months, and 1.43% within ±24 months compared to the method in [37]. Moreover, we further test our MAAL-Net on CNBA dataset, achieving an accuracy of 87.27% within ±6 months and 96.65% within ±12 months, which are only approximately 2% lower than those on the RSNA dataset. The accuracy within ±24 months of the proposed method reaches 99.36%, which is close to 100%. In comparison with the MAE of human experts in clinical bone age diagnosis with a value of 7.32 months [18], the accuracies within ±6 months of MAAL-Net on two datasets are 89.14% and 87.27%, which meet the expectations for clinical bone age diagnosis.

4.4. Analysis of Hyper-Parameters

Ablation experiments are performed on ambiguous labels and multi-branch attention structures. To analyze the effect of hyper-parameter σ, MAAL-Net is tested several times for σ values ranging from 1 to 4. As shown in Table 3, MAAL-Net obtains the highest accuracy with MAE = 4.07 when σ = 3. We can find that a suitable σ can effectively reduce the MAE. For the Gaussian distribution, the area in the horizontal axis interval (−2.58σ, +2.58σ) is 99.730020%. When σ = 3, we obtain 2.58σ = 7.74, which is close to the MAE = 7.32 of human experts in clinical bone age diagnosis using the Greulich and Pyle approach [18]. Therefore, it is a good choice to set σ = 3. As we can see, the lowest MAE performance is simply obtained from the hand image branch which is used as the final prediction output during the test phase.

4.5. Visualization and Interpretation

We further perform a visualization experiment to effectively and intuitively understand the proposed method. Heat maps generated by Grad-CAM [38] on input images are shown in rows 1 and 3 of Figure 3. The peaks of heat maps indicate that the important features are mainly concentrated in the most distinctive parts, such as the carpal and phalanges. For young children, the important features are concentrated on the carpal rather than the phalanges, as shown in columns 1 and 4 in Figure 3. For teenagers and those in late adolescence, important features are concentrated at the carpal and phalanges, and the features of the carpal are stronger than those of the phalanges, as shown in columns 2, 3, 5, and 6 in Figure 3. These findings are consistent with prior knowledge of experts for BAA. Therefore, the proposed method allows us to learn and establish accurate knowledge without prior knowledge of experts. Furthermore, we provide the distribution curves of the predicted output and ground-truth labels as shown in rows 2 and 4 in Figure 3. As we can see, most of the prediction and ground-truth label distribution curves have very strong consistency. important features are concentrated in the wrist and toe bones, but the features of the wrist bones diminish while those of the toe bones increase, as shown in columns 2, 3, 5, and 6 in Figure 3.

5. Conclusions

This study presents a novel automatic BAA method MAAL-Net based on multi-branch attention with ambiguous labels. This method addresses the lack of accurately labeled images with ROI annotations from different races by extracting ambiguous labels using NLP techniques from radiological reports for training. In particular, our method solves the problem that existing models cannot be used for the BAA of Eastern children. The proposed MAAL-Net makes effective use of the ambiguous labels in feature and classifier learning, and the multi-branch attention structure can fully learn global and local features to achieve excellent performance. Compared to methods that require accurately labeled datasets, our approach achieves competitive results. Notably, the accuracy of our method within ±6 months is significantly improved compared to previous studies. These results demonstrate the potential of our model to be applied to the BAA of Chinese children, reducing the workload of experts and providing a reference for routine screening of children’s growth and development. This study also provides a feasible solution for effective medical imaging learning by exploring medical knowledge from radiology reports in the future.

6. Limitations and Future Directions

Although much effort has been made on the automatic BAA system, it still suffers from some limitations to be further investigated.
First, the dataset is annotated at a relatively coarse granularity. The X-ray image in the dataset has only one coarse label and there is no fine-grained label for the ROI to determine bone age. This makes the network produce a deviation when extracting the key features of the ROI, which further leads to a more serious error in the final classification of the bone age. Due to the limitations of data annotation, the model training in this paper is based on the characteristics of the whole hand bone using the GP atlas method. Although the GP atlas is simple, intuitive, and fast, its prediction accuracy of bone age is inferior to the later TW3 method and CHN05 method. In the follow-up study, we will continuously improve the dataset by labeling 13 different ROI of hand bones and giving the development level of each bone part. For the bone age prediction methods, TW3 and CHN05 methods with higher accuracy will be adopted. For the backbone network, we will use the transformer-based DETR [39] (Detection Transformer) to improve the accuracy of hand bone ROI positioning and enhance interpretability. The accuracy of bone age prediction will also be further improved.
Secondly, there is an imbalance in data categories. The experiment in this paper is conducted on hand images of children aged 37–228 months. The large span of bone age and the significant differences in the data volume of each age group, especially the relatively small number of bone age images in late adolescence, lead to the different accuracy of bone age prediction for different age groups in the test phase. In the follow-up study, we consider conducting independent training for different age groups, such as preschool, school-age, and adolescents. Furthermore, new loss functions are supposed to train the network. For example, we will use LMFLOSS [40] function to dynamically consider hard samples and class distribution and alleviate the poor performance caused by data class imbalance.
Finally, there is the limitation of domain shift. The X-ray image of patients’ hands are obtained from different machines and equipment in multiple hospitals. The scanning protocol, filming parameters, shooting angle, and subject group are different, which leads to poor generalization performance in the subsequent tests in different hospitals. In future research, we will focus on the deep feature representation of data to capture more information conducive to the bone age prediction task, select the closest two domain samples, close the same category distance between the two domains, and widen the distance between different categories, to achieve cross-domain similarity and dissimilarity, improve bone age prediction precision, and make it have better generalization ability and stability.

Author Contributions

Conceptualization, B.H. and Z.X.; methodology, B.H. and Z.X.; software, B.H. and Y.C.; writing—original draft preparation, B.H. and Z.X.; writing—review and editing, B.H. and Z.X.; validation, D.Z.; visualization, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

https://www.kaggle.com/kmader/rsna-bone-age (accessed on 21 March 2023).

Conflicts of Interest

The authors declare that they have no conflict of competing interests.

References

  1. Gertych, A.; Zhang, A.; Sayre, J.; Pospiech-Kurkowska, S.; Huang, H.K. Bone age assessment of children using a digital hand atlas. Comput. Med. Imaging Graph. 2007, 31, 322–331. [Google Scholar] [CrossRef]
  2. Teele, R. Assessment of Skeletal Maturity and Prediction of Adult Height. J. Paediatr. Child Health 2003, 39, 322–323. [Google Scholar] [CrossRef]
  3. Greulich, W.W.; Pyle, S.I. Radiographic Atlas of Skeletal Development of the Hand and Wrist. Am. J. Med. Sci. 1959, 238, 393. [Google Scholar] [CrossRef]
  4. Tanner, J.M.; Whitehouse, R.; Cameron, N.; Marshall, W.; Healy, M. Assessment of Skeletal Maturity and Prediction of Adult Height (TW2 Method). Am. J. Hum. Biol. 1976, 14, 788–798. [Google Scholar]
  5. Fischbach, F.; Knollmann, F.; Griesshaber, V.; Freund, T.; Akkol, E.; Felix, R. Detection of pulmonary nodules by multislice computed tomography: Improved detection rate with reduced slice thickness. Eur. Radiol. 2003, 13, 2378–2383. [Google Scholar] [CrossRef]
  6. Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilcus, S.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.; Shpanskaya, K.; et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. In Proceedings of the 2019 AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 590–597. [Google Scholar]
  7. Ji, Y.L.; Kim, J.S.; Kim, T.K.; Kim, Y.S. Detection and classification of intracranial haemorrhage on CT images using a novel deep-learning algorithm. Sci. Rep. 2020, 10, 20546. [Google Scholar]
  8. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  11. Jia, D.; Wei, D.; Socher, R.; Li, L.J.; Kai, L.; Li, F.F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  12. Pooch, E.H.P.; Ballester, P.L.; Barros, R.C. Can we trust deep learning models diagnosis? The impact of domain shift in chest radiograph classification. arXiv 2019, arXiv:1909.01940. [Google Scholar]
  13. Moreno-Torresa, J.G.; Raeder, T.; Alaiz-Rodríguez, R.; Chawla, N.V.; Herreraa, F. A unifying view on dataset shift in classification. Pattern Recognit. 2012, 45, 521–530. [Google Scholar] [CrossRef]
  14. Lin, S.; Ying, K.; Chen, S.; Lee, Z. Particle swarm optimization for parameter determination and feature selection of support vector machines. Expert Syst. Appl. 2008, 35, 1817–1824. [Google Scholar] [CrossRef]
  15. Harmsen, M.; Fischer, B.; Schramm, H.; Seidl, T.; Deserno, T.M. Support Vector Machine Classification Based on Correlation Prototypes Applied to Bone Age Assessment. IEEE J. Biomed. Health Inform. 2013, 17, 190–197. [Google Scholar] [CrossRef] [PubMed]
  16. Somkantha, K.; Theera-Umpon, N.; Auephanwiriyakul, S. Bone age assessment in young children using automatic carpal bone feature extraction and support vector regression. J. Digit. Imaging 2011, 24, 1044–1058. [Google Scholar] [CrossRef]
  17. Thodberg, H.H.; Kreiborg, S.; Juul, A.; Pedersen, K.D. The BoneXpert Method for Automated Determination of Skeletal Maturity. IEEE Trans. Med. Imaging 2009, 28, 52–66. [Google Scholar] [CrossRef]
  18. Larson, D.B.; Chen, M.C.; Lungren, M.P.; Halabi, S.S.; Stence, N.V.; Langlotz, C.P. Performance of a Deep-Learning Neural Network Model in Assessing Skeletal Maturity on Pediatric Hand Radiographs. Radiology 2017, 287, 313–322. [Google Scholar] [CrossRef] [PubMed]
  19. Spampinatoa, C.; Palazzoa, S.; Giordanoa, D.; Aldinuccib, M.; Leonardic, R. Deep learning for automated skeletal bone age assessment in X-ray images. Med. Image Anal. 2017, 36, 41–51. [Google Scholar] [CrossRef]
  20. González, C.; Escobar, M.; Daza, L.; Torres, F.; Triana, G.; Arbeláez, P. SIMBA: Specific Identity Markers for Bone Age Assessment. In Proceedings of the 2020 Medical Image Computing and Computer Assisted Intervention, Lima, Peru, 4–8 October 2020. [Google Scholar]
  21. Escobar, M.; González, C.; Torres, F.; Daza, L.; Triana, G.; Arbeláez, P. Hand Pose Estimation for Pediatric Bone Age Assessment. In Proceedings of the 2019 Medical Image Computing and Computer Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  22. Wang, F.; Gu, X.; Chen, S.; Liu, Y.; Shen, Q.; Pan, H.; Jin, Z. Artificial intelligence system can achieve comparable results to experts for bone age assessment of Chinese children with abnormal growth and development. PeerJ 2020, 8, e8854. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, D.; Zhang, D.; Ding, J.; Wang, J. Improve bone age assessment by learning from anatomical local regions. arXiv 2020, arXiv:2005.13452. [Google Scholar]
  24. Chandran, J.J.G.; Karthick, R.; Rajagopal, R.; Meenalochini, P. Dual-Channel Capsule Generative Adversarial Network Optimized with Golden Eagle Optimization for Pediatric Bone Age Assessment from Hand X-Ray Image. Int. J. Pattern Recognit. Artif. Intell. 2023, 37, 2354001. [Google Scholar] [CrossRef]
  25. Geng, X. Label Distribution Learning. IEEE Trans. Knowl. Data Eng. 2016, 28, 1734–1748. [Google Scholar] [CrossRef]
  26. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  27. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  28. Neubeck, A.; Gool, L. Efficient Non-Maximum Suppression. In Proceedings of the 18th International Conference on Pattern Recognition, Hong Kong, China, 20–24 August 2006; pp. 850–855. [Google Scholar]
  29. Laine, S.; Aila, T. Temporal Ensembling for Semi-Supervised Learning. In Proceedings of the 2016 International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  30. Halabi, S.S.; Prevedello, L.M.; Kalpathy-Cramer, J.; Mamonov, A.B.; Bilbily, A.; Cicero, M.; Pan, I.; Pereira, L.A.; Sousa, R.T.; Abdala, N.; et al. The RSNA Pediatric Bone Age Machine Learning Challenge. Radiology 2019, 290, 498–503. [Google Scholar] [CrossRef]
  31. Lee, H.; Tajmir, S.; Lee, J.; Zissen, M.; Yeshiwas, B.A.; Alkasab, T.K.; Choy, G.; Do, S. Fully Automated Deep Learning System for Bone Age Assessment. J. Digit. Imaging 2017, 30, 427–441. [Google Scholar] [CrossRef] [PubMed]
  32. Iglovikov, V.L.; Rakhlin, A.; Kalinin, A.A.; Shvets, A.A. Paediatric Bone Age Assessment Using Deep Convolutional Neural Networks. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018. [Google Scholar]
  33. Liu, C.; Xie, H.; Liu, Y.; Zha, Z.; Lin, F.; Zhang, Y. Extract Bone Parts Without Human Prior: End-to-end Convolutional Neural Network for Pediatric Bone Age Assessment. In Proceedings of the 2019 Medical Image Computing and Computer Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  34. Steenkiste, T.V.; Ruyssinck, J.; Janssens, O.; Vandersmissen, B.; Vandecasteele, F.; Devolder, P.; Achten, E.; Hoecke, S.V.; Deschrijver, D.; Dhaene, T. Automated Assessment of Bone Age Using Deep Learning and Gaussian Process Regression. In Proceedings of the 2018 IEEE Engineering in Medicine and Biology Society, Honolulu, HI, USA, 18–21 July 2018. [Google Scholar]
  35. Pan, X.; Zhao, Y.; Chen, H.; Wei, D.; Zhao, C.; Wei, Z.; Soliman, A. Fully Automated Bone Age Assessment on Large-Scale Hand X-Ray Dataset. Int. J. Biomed. Imaging 2020, 2020, 8460493. [Google Scholar] [CrossRef]
  36. Chen, C.; Chen, Z.; Jin, X.; Li, L.; Speier, W.; Arnold, C.W. Attention-Guided Discriminative Region Localization for Bone Age Assessment. IEEE J. Biomed. Health Inform. 2021, 26, 1208–1218. [Google Scholar] [CrossRef] [PubMed]
  37. Hao, P.; Ye, T.; Xie, X.; Wu, F.; Ding, W.; Zuo, W.; Chen, W.; Wu, J.; Luo, X. Radiographs and texts fusion learning based deep networks for skeletal bone age assessment. Multimed. Tools Appl. 2020, 80, 16347–16366. [Google Scholar] [CrossRef]
  38. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  39. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  40. Sadi, A.A.; Chowdhury, L.; Jahan, N.; Rafi, M.N.S.; Chowdhury, R.; Khan, F.A.; Mohammed, N. LMFLOSS: A Hybrid Loss for Imbalanced Medical Image Classification. arXiv 2022, arXiv:2212.12741. [Google Scholar]
Figure 1. Structure of Multi-branch Attention Learning with Ambiguous Label Network (MAAL-Net).
Figure 1. Structure of Multi-branch Attention Learning with Ambiguous Label Network (MAAL-Net).
Sensors 23 04834 g001
Figure 2. Bone age distributions.
Figure 2. Bone age distributions.
Sensors 23 04834 g002
Figure 3. Visualization of bone age prediction results. Rows 1 and 3 show the attention heat maps of the RSNA dataset and CNBA dataset, respectively. Rows 2 and 4 show the bone age prediction results of X-ray images from the RSNA dataset and CNBA dataset, respectively. The solid line represents the ground-truth label transformation into Gaussian distribution, while the dotted line represents the label distribution output by the model prediction.
Figure 3. Visualization of bone age prediction results. Rows 1 and 3 show the attention heat maps of the RSNA dataset and CNBA dataset, respectively. Rows 2 and 4 show the bone age prediction results of X-ray images from the RSNA dataset and CNBA dataset, respectively. The solid line represents the ground-truth label transformation into Gaussian distribution, while the dotted line represents the label distribution output by the model prediction.
Sensors 23 04834 g003
Table 1. Mean Absolute Error (MAE) results of different methods.
Table 1. Mean Absolute Error (MAE) results of different methods.
MethodsDatasetRoI AnnotationAmbiguous LabelMAE (M)
Human Expert et al. [18]RSNA--7.32
Larson et al. [18]RSNANONO6.24
Spampinato et al. [19]PrivateNONO9.12
Iglovikoc et al. [32]RSNAYESNO4.97
Tom et al. [34]RSNANONO6.8
BoNet [21]RSNAYESNO4.14
AR-CNN [33]RSNAYESNO4.38
Chen et al. [36]RSNANONO4.7
Pan et al. [35]RSNANONO7.35
SIMBA [27]PrivateNONO5.47
MAAL-Net [ours]RSNANOYES4.09
Table 2. Accuracy results of different methods.
Table 2. Accuracy results of different methods.
MethodsDataset±6 Months±12 Months±24 Months
Hao et al. [37]RSNA55.28%91.09%98.50%
MAAL-Net [ours]RSNA89.14%98.24%99.93%
CNBA87.27%96.65%99.36%
Table 3. Ablation study on the construction of ambiguous labels and multi-branch attention on RSNA dataset.
Table 3. Ablation study on the construction of ambiguous labels and multi-branch attention on RSNA dataset.
Hyper-ParameterRaw Image BranchHand Image BranchBone Part Images Branch
σ = 15.925.17.54
σ = 25.724.876.53
σ = 34.864.075.68
σ = 45.294.95.96
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, B.; Xu, Z.; Zhou, D.; Chen, Y. Multi-Branch Attention Learning for Bone Age Assessment with Ambiguous Label. Sensors 2023, 23, 4834. https://doi.org/10.3390/s23104834

AMA Style

He B, Xu Z, Zhou D, Chen Y. Multi-Branch Attention Learning for Bone Age Assessment with Ambiguous Label. Sensors. 2023; 23(10):4834. https://doi.org/10.3390/s23104834

Chicago/Turabian Style

He, Bishi, Zhe Xu, Dong Zhou, and Yuanjiao Chen. 2023. "Multi-Branch Attention Learning for Bone Age Assessment with Ambiguous Label" Sensors 23, no. 10: 4834. https://doi.org/10.3390/s23104834

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop