Next Article in Journal
MRI of Implantation Sites Using Parallel Transmission of an Optimized Radiofrequency Excitation Vector
Next Article in Special Issue
Animal Models and Their Role in Imaging-Assisted Co-Clinical Trials
Previous Article in Journal
Two-Dimensional Post-Traumatic Measurements of Orbital Floor Blowout Fractures Underestimate Defect Sizes Compared to Three-Dimensional Approaches
Previous Article in Special Issue
Repeatability of Quantitative Magnetic Resonance Imaging Biomarkers in the Tibia Bone Marrow of a Murine Myelofibrosis Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Repeatability of Mouse Tibia Volume Segmentation in Murine Myelofibrosis Model Using Deep Learning

Department of Radiology, University of Michigan, Ann Arbor, MI 48109, USA
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Tomography 2023, 9(2), 589-602; https://doi.org/10.3390/tomography9020048
Submission received: 11 January 2023 / Revised: 2 March 2023 / Accepted: 3 March 2023 / Published: 7 March 2023

Abstract

:
A murine model of myelofibrosis in tibia was used in a co-clinical trial to evaluate segmentation methods for application of image-based biomarkers to assess disease status. The dataset (32 mice with 157 3D MRI scans including 49 test–retest pairs scanned on consecutive days) was split into approximately 70% training, 10% validation, and 20% test subsets. Two expert annotators (EA1 and EA2) performed manual segmentations of the mouse tibia (EA1: all data; EA2: test and validation). Attention U-net (A-U-net) model performance was assessed for accuracy with respect to EA1 reference using the average Jaccard index (AJI), volume intersection ratio (AVI), volume error (AVE), and Hausdorff distance (AHD) for four training scenarios: full training, two half-splits, and a single-mouse subsets. The repeatability of computer versus expert segmentations for tibia volume of test–retest pairs was assessed by within-subject coefficient of variance (%wCV). A-U-net models trained on full and half-split training sets achieved similar average accuracy (with respect to EA1 annotations) for test set: AJI = 83–84%, AVI = 89–90%, AVE = 2–3%, and AHD = 0.5 mm–0.7 mm, exceeding EA2 accuracy: AJ = 81%, AVI = 83%, AVE = 14%, and AHD = 0.3 mm. The A-U-net model repeatability wCV [95% CI]: 3 [2, 5]% was notably better than that of expert annotators EA1: 5 [4, 9]% and EA2: 8 [6, 13]%. The developed deep learning model effectively automates murine bone marrow segmentation with accuracy comparable to human annotators and substantially improved repeatability.

1. Introduction

Myelofibrosis (MF) is a chronic, ultimately fatal myeloproliferative neoplasm caused by genetic mutations in hematopoietic stem cells that cause systemic inflammation and progressive fibrosis, disrupting normal architecture and composition of the bone marrow [1,2]. Bone marrow biopsy remains the standard of care to assess the primary site of disease in MF [2,3]. These biopsies are painful and inherently suffer from sampling error, as the technique only analyzes millimeter amounts of tissue from one anatomic site (iliac crest) and do not assess heterogeneity of disease [4]. Development of quantitative imaging biomarkers (QIBs) of MF would facilitate comprehensive monitoring of disease progression and therapy response in patients in place of needle biopsies.
The goal of our ongoing co-clinical MF study [5,6] is to develop robust imaging protocols and analysis methods for quantitative imaging biomarkers of myelofibrosis. Investigation of the murine model of MF disease for preclinical testing of new therapeutic interventions [7] allows direct exploration of correlations between MRI and bone marrow pathology as full bone marrow can be extracted and studied under microscope [6]. To determine quantitative thresholds for significant changes in MRI biomarkers related to MF pathology, their measurement errors must be evaluated from repeatability studies [8,9] that include all the acquisition and analysis steps. Establishing QIB confidence intervals allows detection of biological changes above the random measurement errors. Currently, “test–retest” experiments are standard means to estimate QIB repeatability in clinical oncology imaging studies, e.g., of lung, prostate, and breast cancer [10,11,12,13]. Following this practice, test–retest scans were collected for the preclinical arm of the MF study to assess bone marrow QIB repeatability [14].
The first essential step of the MF QIB analysis is segmenting the bone marrow within which QIB is measured. Manual segmentation by human operators is typically a tedious and time-consuming procedure and is prone to inter- and intra-observer variability [15,16]. Recently developed deep learning (DL) approaches show promise in both saving human effort and improving image segmentation accuracy and reproducibility [17,18]. One of the robust tools for image segmentation in the field of biomedical imaging is a U-Net based on deep convolutional neural network (DCNN) [19,20,21]. Convolution-based segmentation models take advantage of the spatial nature of images and learn well from relatively small training datasets common for medical imaging due to a shortage of available comprehensive expert annotations [18,21]. Thus, DCNNs with U-Net-based architectures currently are widely used for medical image segmentation [21,22]. In the current study, we evaluated the Attention U-Net (A-U-Net) consisting of a U-Net incorporated with attention gates [23] that can automatically learn to focus on important image features.
Key aspects where DL tools could help our ongoing co-clinical MF project are (1) to improve segmentation efficiency through automation of an otherwise tedious task for human experts, and (2) to improve precision of QIB by producing more repeatable segmentation results that serve as key input to QIB analysis. To explore the capability of DL tools for the above tasks, our current study focuses on assessing the A-U-Net model segmentation repeatability and relative accuracy compared to human expert segmentation of murine tibia, as well as the estimation of practical training size that can provide robust model performance.

2. Materials and Methods

The current study sought to determine whether the A-U-Net DCNN model segmentation of mice tibia would be more repeatable than manual segmentation without compromised accuracy. We performed test–retest repeatability and relative accuracy analysis on tibial bone volumes (scanned on two consecutive days) obtained by a deep learning (DL) segmentation model with respect to two human experts. This study design assumes that the tibia volume of a given mouse does not change significantly over the period of 1 day so that repeatability of the segmentation process can be assessed objectively without knowledge of the true tibia volume.

2.1. Experimental Design and Dataset

The studied dataset consisted of 3D Gradient echo (3D-GRE) MR images for 32 mice (26 diseased and 6 normal controls) with a total of 157 scans. The scans were acquired over a three-month period after induction of myelofibrosis by bone-marrow ablation for female mice with JK1 mutation. Forty-nine test–retest scan pairs were collected on consecutive days to assess repeatability [8,9]. The MRI acquisition parameters are detailed in Appendix A. Conversion from the acquired scanner native 3D magnitude-valued 2dseq-image format to the meta-image header (MHD) ITK format was performed using custom routines in MATLAB (R2016b+, Mathworks, Natick MA, USA). The resulting MHD 3D volumes had a dimension of 128 × 64 × 256 voxels with voxel size of 0.09 × 0.075 × 0.094 mm3 (voxel volume of 6.35 × 10−4 mm3) and gray-scale intensity bit-depth of 14; indexing from each dimension provided different cross-sectional views (Figure 1, inset).
The MHD images were used for DL model training and data analysis. The experimental design is shown in Figure 1. The available dataset was split into training (18 diseased mice, five controls: 107 scans), validation (two diseased mice, one control: 17 scans), and test (six diseased mice: 33 scans) subsets to approximately balance by scan number (70%/10%/20%). The validation and test subsets were not altered during evaluation. Due to the small number of normal mice and our intended application for disease segmentation, their scans were included only in training (21 scans) and validation (six scans) sets. The full training set included 32 test–retest pairs, validation set contained four pairs, and test set had 13 pairs. To assess model sensitivity to the training set (TS) size, the full TS was split into three training subsets (Figure 1). Training split 1 (TS1) and training split 2 (TS2) each contained roughly half of the training data. Each split was balanced on the basis of test–retest pairs. The training single mouse (TSM) subset consisted of a single mouse randomly selected with a long time-series of (six biweekly scans), with four test–retest pairs.
Two expert annotators provided the tibia segmentation contours for the test and validation subsets. The masks contoured by Expert Annotator 1 (EA1) with 3 years of experience were used for training and as a reference for model accuracy assessment. Expert Annotator 2 (EA2) was instructed by EA1 (on a single-mouse example) to perform segmentations of the test and validation subsets. Scan and rescan annotations were performed independently. Prior to instruction, EA2 also dispatched the (unbiased) data subset selection for DL model training, validation, and testing. The reference tibia contours were drawn on the coronal view typically spanning 15–30 slices; thus, the coronal view was selected for segmentation by the DL model. The dimension of each grayscale MHD image in coronal view was 128 × 256, thus providing 64 2D images per mouse scan.

2.2. DL Model Architecture

We selected the A-U-net DL model for the tibia segmentation based on improved performance observed in our preliminary comparison study [24]. Our U-Net was implemented using Pytorch library (version 1.13.0) The base U-Net model architecture is detailed in [19,25]. Attention U-Net adds attention gates to the U-Net architecture [19,23] to increase the weights for automatically detected important image features, as illustrated in Figure 2. Each MRI slice was treated as an independent image, and the batch size (N) was randomly sampled from all available slices. The size of the input was (N,1,H,W) for a batch size of N, image height H, and width W. Attention U-Net consists of four encoder blocks and four decoder blocks (Figure 2A). The encoder network (contracting path) preserves the spatial dimensions and doubles the number of filters after each stage whereas the decoder network (expanding path) doubles the spatial dimensions and halves the number of filters.
The attention gate (Figure 2B) was described in [23]. Briefly, it takes the input signal (shape: (N,F/2,2H,2W)) and gating signal (shape: (N,F,H,W)) from two different stages, where the input signal and the gating signal have different sizes. The convolutions applied to the input signal (filter size = 1 × 1, stride 2, output size = F) and the gating signal (filter size = 1 × 1 and stride 1, output size = F) equalize their sizes, which are then added, and the rectified linear unit (ReLU) activation function is applied, followed by convolution (filter size = 1 × 1, stride =1, pad = 1, output size = 1) and sigmoid activation. Next, bilinear interpolation is applied to double the height and width to match that of the input signal, producing an attention coefficient array (N,1,2H,2W). The dot product of the attention coefficients with the input signal is the output from the attention gate that emphasizes the important features for image segmentation in the expanding path of the network (Figure 2A).
We experimented with training the Attention U-Net on the complete subset (Figure 1, TS), 50% splits (Figure 1, TS1 and TS2), and a single mouse subset (Figure 1, TSM) with respect to the EA1 segmentations as reference. The model was trained to minimize the binary cross-entropy loss. We fine-tuned the hyperparameters and selected the best model using the validation dataset for the different training scenarios. The model training typically took up to 120 epochs with 10,500 iterations for TS and batch size of 40 (6 h for TS, 3 h for TS1 or TS2, and 20 min for TSM using NVIDIA 1080Ti.) The best models were then deployed to the held-out test set, and the automated segmentations were compared to the manual EA1 and EA2 segmentations.

2.3. Accuracy and Repeatability Evaluation

The DL model performance was assessed by segmentation accuracy with respect to EA1 reference and repeatability from test–retest experiments. Since true tibia volumes were unknown, only “relative accuracy” was assessed by this procedure. For an evaluation of the relative model accuracy, we used four commonly used performance metrics [19], namely, average Jaccard index (AJI), average volume intersection ratio (AVI), average volume error (AVE), and average Hausdorff distance (AHD) to compare the performance of the models and expert annotators (EA1 and EA2). The metrics are defined in Appendix B. The average of a given metric is the average over the entire set. Student’s t-test with Bonferroni correction was applied to compare accuracy metrics (with respect to EA1 reference) of test-set mouse segmentations for different training scenarios (Figure 1, full training set, TS1, TS2, and TSM), and with EA2. The differences were considered significant for p-value < 0.009.
The relative accuracy of the model volume segmentations for the test subset (Figure 1) was also assessed by comparison to the EA1 reference via Bland–Altman (BA) analysis [9] and the agreement was quantified by Pearson correlation, R. BA analysis was also used to evaluate test–retest repeatability for volume segmentations by individual expert annotators and DL models trained on different data subsets. Bland–Altman plots provided a graphic view of agreement.
For the scan pairs acquired on consecutive days, the volume of the mouse tibia is not expected to change significantly; thus, the differences in the segmented volumes should ideally be zero. The repeatability and comparability of volumes estimated from the test–retest pairs were assessed using the within-subject coefficient of variation (wCV%) and BA analysis for limits of agreement (LOA) [9,26]. wCV is a relative (dimensionless) repeatability metric with confidence intervals (CIs) defined in Appendix B. The bias and LOA together determine the similarity between the test and retest volume pair. Small bias and narrow LOA indicate that the two measurements are essentially equivalent.

3. Results

3.1. DL Model Segmenation Accuracy with Respect to Reference

An example of the tibia segmentation from the validation set is presented in Figure 3. A darker tibia bone boundary is clearly visible. All the models accurately predicted the tibia bone contour with very minimal variations from the expert annotations even in regions of signal heterogeneity along the tibia length (Figure 3). The change in the tibia volumes over a period of 6 weeks (Figure 3E) was consistent with minor changes in the contours.
As expected, the DL model performance metrics (Table 1) reflect the highest relative accuracy (AJI: 86%–88%; AVI: 93%–97%) for the training sets. As the training set size is reduced, the standard deviations (SDs) get wider (e.g., from 5% to 10%) for the test subset segmentations.
However, the average performance metrics are sufficiently close (within SD) for the TS1, TS2, and full training subsets, and similar accuracy is observed for the test and validation subsets when the models were trained on half of the training data. Compared to training with the full set, the accuracies and performance metrics for the test subset were only slightly lower (<1.5%) when the models were trained on TS1 or TS2, which was not significantly different after Bonferroni correction (p > 0.015). For the test subset, the relative accuracy of DL model trained on a single mouse was significantly lower compared to the other training scenarios. The AHDs were below 1 mm for all training scenarios, increasing with decreasing training set size.
The lower value of AJI% for EA2 vs. EA1 reflects the interobserver variability in the way EA1 and EA2 generate the masks. There is large over-segmentation by EA2 relative to EA1 as observed from the AVEs of 14–18%. The interobserver variability in mouse tibia segmentation by the human experts is much larger than that of the A-U-Net models relative to EA1 (<3%). For the test set, the relative accuracy of EA2 segmentations with respect to the EA1 reference (AJI = 81%, AVI = 83%, AVE = 14%) was notably below that achieved by A-U-net models trained on full or TS1 and TS2 subsets (AJI = 82.5–83.5%, AVI = 89–90%, AVE = 2–3%), with significant improvement for AVI (6%) and AVE (−11%).
In the test example shown in Figure 4, the DL model segmentations derived after training with full and half datasets (Figure 1, TS1 and TS2) apparently produced slightly more accurate tibia segmentation at the distal end compared to somewhat under-segmented EA1 “reference” (Figure 4B, yellow arrow). Since expert segmentation may have variabilities, the agreement or disagreement with respect to the reference annotations may not reflect the true model accuracy (Table 1). The predictions made by all the models (Attention U-Net trained on multiple mice) are fairly robust except for the one trained on a single mouse (Figure 4G). The predicted tibia mask by the model trained on the TSM subset misses a large portion of bone contour near the knee (indicated by the yellow arrow in Figure 4G). Thus, the model learned from a small training set shows declining performance and is more susceptible to errors in reference annotations of the training samples. The small data size fails to include all the diversity and features that would be encountered while making the test predictions.
Figure 5 further summarizes the relative accuracy of volume segmentations in the test subset with respect to the EA1 reference for different A-U-Net training scenarios. EA2 segmentations show systematic positive bias ~1 mm3 (consistent with tendency to over-segment) while the A-U-Net segmentations are essentially free of bias, which indicates that the models learn to segment in a similar way as EA1 as they were all trained with the EA1 annotations. Except for single-mouse training (TSM) (dispersion ~2 mm3, R ~0.74, all training scenarios are in good agreement with EA1 reference (dispersion <1 mm3 and model-generated volumes highly correlated to the reference: R ~0.92 to 0.95). The DL models are also highly correlated among themselves (R ~0.98 to 0.99). The notable decline in relative model precision for TSM confirms the insufficient training set size (Figure 1, TSM).

3.2. DL Model Segmenation Precision

Table 2 further summarizes the observed repeatability trends in volume estimations for the test–retest pairs in all data subsets and volumes. For single-mouse training and all validation results that included only four test–retest pairs (Figure 1), the CI range is quite broad (from 6% to 35%), likely resulting from insufficient training to handle the large variations in image properties of unknown validation and test cases and, thus, causing inconsistent segmentation of the tibia volumes. As expected, for the 16–32 test–retest pairs in the training subsets, the model-derived tibia volume repeatability is close to that of the EA1 reference with an average wCV = 7.7% and CI = 5.5–12.3%, as the A-U-Net models were specifically trained to follow the EA1 segmentations.
Similar precision performance is observed for EA2 on the test set with wCV = 8%, while EA1 with wCV = 5.3% is notably better, but there is substantial CI overlap (3.8–8.7% versus 5.7–13.2%), reflecting inter- and intra-observer variabilities for manually segmented tibia volumes. In contrast, for DL models trained on the full set, TS1 or TS2, the average repeatability (wCV ~3%) with tight CI = 1.9–5.3% is about twofold better than that for single-mouse training or for EA1 and EA2 segmentations of the test–retest pairs in the test set.
The repeatability of tibia volume measurements for all DL models and expert annotators is compared for the test–retest pairs in the test subset in Figure 6. The models show negligible (positive) volume bias between the scan and rescan segmentations, while EA2 shows notable positive bias (0.4 mm3), consistent with the tendency of EA2 to slightly over-segment for the repeated scan (similar to over-segmentation relative to EA1 as observed in Figure 5). The LOAs for tibia volumes generated by DL models trained on all or half the data are small with the overall precision within 0.5 mm3. The precision of EA1 is about half of DL model with the spread up to 1 mm3. For the EA2 and DL model trained on a single mouse, the precision is threefold lower, with an overall spread of about 1.5 mm3.

4. Discussion

The main finding of our study is that DL-based segmentation shows strong promise in improving repeatability of tibia volume measurements while the accuracy is comparable to manual annotations. For the test subset, the accuracy of DL segmentation trained on more than 10 animals (>52 scans) notably exceeded that of the second expert annotator (by 6% for volume intersect ratio and by −11% for volume error), with repeatability errors reduced to 3% from 5%–8% (observed for manual segmentations). Thus, the developed model has strong potential to enhance precision of the corresponding volume-based bone marrow QIBs of myelofibrosis [6] with substantial saving of human effort. The utilized test–retest performance metrics is generally more objective than relative accuracy assessment and would be useful for routine evaluation of DL model-based segmentation precision.
The variability in volumes from manual segmentation is a known major factor that limits precision in quantitative imaging [12,13,27]. The manual image segmentation is a tedious time-consuming task with repeatability dependent on the level of experience, attention span and fatigue of an annotator. Variabilities among the experts are also known to contribute to reproducibility errors [15,16]. Since human annotations are generally not free from subjective judgments, only relative segmentation accuracy can be assessed for the DL models with respect to the expert-provided reference. In fact, we noticed that the DL model trained on multiple mice can provide more accurate segmentations compared to DL model trained on a single mouse, while lower performance measures could just indicate discrepancy with the expert reference. In contrast, the segmentation repeatability is a more objective measure of model performance since the tibia volumes should not change between test and retest scans.
For clinical QIB precision assessments, QIBA recommends a test–retest sample size of >35 [8] to evaluate wCV with nominal CIs. Our preclinical validation sample size was insufficient for confident wCV measurement, but the samples in the test subset provided moderate CIs that allowed confident detection of twofold improved repeatability of DL model-based mouse tibia segmentations (3%) compared to expert annotations (5.3% to 8%) with relatively high accuracy (>82%) when the model was trained on the dataset of the size exceeding test dataset size. Apparently, training on 53–54 scans for 11–12 animals enabled relative accuracy and repeatability for test set segmentation equivalent (within confidence intervals) to training on 107 scans for 23 animals. Small image datasets with limited expert annotations are typical for biomedical image segmentation applications [28,29]. The training sample size required for robust model training is expected to depend on the complexity of the task and the DL architectures used. Therefore, estimating reasonable size for DL model training for a given task is of practical value to save human effort and improve segmentation consistency [30,31].
To date DL segmentation of murine MRI data has predominantly focused on application to soft-tissue regions and lesions [32,33]. In contrast to prior DL segmentation studies of murine bone primarily performed for microcomputed tomography (µCT) images [30,34,35], our current study provides the first example of DL bone segmentation in high-resolution MRI data. We studied the use of 2D images and the dependence on training sample sizes for training A-U-Net models. The relative accuracy of 80–90% and AHD <1 mm achieved for bone segmentation are similar to those reported for the µCT studies [30,34,35]. Although prior DL segmentation work compared relative model performance for multiple annotators [34] and different CT scanners [35], none used scan–rescan analysis for objective precision assessment as reported in our study.
Our study had limitations. First, the test and training subsets (of relatively small sizes) came from the same dataset acquired on a single scanner with fixed scan protocol for a single mouse mutation model; second, the reference annotations were provided by a single expert. These would likely limit direct generalization of the model for different acquisition protocols and should be tested independently for different murine models. However, the implementation of the automated segmentation provided valuable insights into practical DL model-based segmentation workflow, which could be utilized for other quantitative MRI acquisition protocols and image datasets.
The mouse tibia anatomy is largely consistent among the animals, allowing the use of highly uniform imaging protocols. Thus, mouse tibia segmentation may represent a relatively straightforward task for DCNN model, providing robust example for development of performance evaluation workflow. Apparently, the Attention U-Net trained on more than 50 scans successfully handled the challenges of both heterogeneous intensities and small murine bone sizes. The developed DL model for murine tibia bone marrow segmentation can be directly applied to future mouse scans acquired with similar MRI protocols. In future work, iterative semi-supervised DL training could be implemented for murine bone segmentation. In such an approach, manual segmentation is performed for a small subset of mouse scans for model training, followed by automatic segmentation and quality check for a small batch of new cases and adding manually corrected (challenging) cases to retrain the model on the enlarged annotated set. The process can be iterated for a number of times until the performance of the trained model is acceptable. This approach may substantially reduce the effort of manual segmentation.
Our upcoming myelofibrosis studies will focus on generalizing the DL model workflow for human bone marrow segmentation. The human bone sizes are much larger, and expert annotations are even more time-consuming. Additionally, there is intrinsically higher variability of orientations, anatomic details, and scan parameters compared to mice, which might potentially require a larger training dataset and or transfer learning. Overall, improved processing throughput is vital for development, validation, and implementation of MRI-based quantitative biomarkers for advancing experimental therapeutics in myeloproliferative neoplasm mouse models with the overarching goal of improving patient outcomes.

5. Conclusions

DL-based segmentation shows strong promise of improved repeatability of murine tibia volume measurements from MRI scans with accuracy comparable to manual annotations and μCT. Attention U-Net was found to be an efficient alternative to human expert segmentations of tibia, providing high accuracy and precision for quantitative imaging of myelofibrosis.

Author Contributions

Conceptualization, T.L.C., D.M. and L.M.H.; methodology, L.M.H., H.-P.C. and D.M.; software, A.K. and L.M.H.; validation, A.K., R.F.M. and H.T.; formal analysis, A.K. and H.T.; investigation, K.H.; resources, B.D.R. and K.H.; data curation, K.H. and R.F.M.; writing—original draft preparation, A.K., R.F.M., H.T. and D.M.; writing—review and editing, T.L.C., B.D.R., K.H., H.-P.C. and L.M.H.; visualization, A.K., R.F.M. and D.M.; supervision, T.L.C., D.M. and L.M.H.; project administration, B.D.R. and T.L.C.; funding acquisition, B.D.R. and T.L.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Institute of Health grant numbers R01CA238023, U24CA237683, R01CA190299, U01-CA232931, and R35CA197701.

Institutional Review Board Statement

The animal study protocol was approved by the Institutional Ethics Committee of University of Michigan (protocol code 00006795 approval date:01/21/2016).

Informed Consent Statement

Not applicable.

Data Availability Statement

The mouse MHD image and annotation collection and tibia volume tables are available upon request.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. MRI Acquisition Parameters

All mice were scanned using a fixed 3D FLASH coronal-plane imaging protocol on 7 T 30 cm bore Agilent system magnet with Small Receive CryoProbeTM4-Element Array RF Coil Kit, cryogenically cooled to 20–30 K. The anesthetized mouse leg was held in place between a 3D-printed, leg-shaped mold on posterior side and the CryoProbeTM on anterior side for a 15 min scan. The imaging protocol parameters were defined in Bruker BioSpec® MRI Console Paravision 7.0.0 software and included single-echo; TR/TE = 111 ms/2.99 ms; flash-spoiling; flip angle = 9°; 1 NSA; imaging matrix size: 256 × 128 × 64; voxel size: 0.09 × 0.075 × 0.094 mm3; field of view (FOV): 23 × 9.6 × 6 mm3.

Appendix B. Performance metrics definitions

Jaccard index: It is defined as the ratio of the intersection of the segmentation volumes divided by the union of the volumes of the two 3D masks. The Jaccard index for a single scan is calculated as
Jaccard   Index = V h V c V h V c ,
where Vh and Vc are the volumes of the segmentation mask drawn by human as reference standard and predicted by computer, respectively, for the same scan.
Volume intersection ratio: The volume intersection ratio for the segmentations of a single scan is calculated as
Volume   Intersection   Ratio = V h V c V h ,
where Vh and Vc are defined as above.
Volume error (AVE): The volume error for the segmentations of a single scan is calculated as
volume   error = V h V c V h ,
where Vh and Vc are defined as above. A positive value of volume error indicates under-segmentation, and a negative value indicates over-segmentation by the computer.
Hausdorff distance (AHD):
HD = m a x {   m a x x r 1 {   m i n y r 2 {   d ( x , y )   } } ,   m a x y r 2 {   m i n x r 1 {   d ( x , y )   } }   } .
The Hausdorff distance measures the maximum distances between the closest points of two segmentation contours, where r1 and r2 are set of points in 3D of the two contours respectively, and d ( x , y ) is the Euclidean distance.
Within-subject coefficient of variance (wCV):
The sum and difference between the paired observations are conveniently used to calculate pairwise mean M i and pairwise variance V i for each i t h test–retest pair as follows:
M i = ( V 1 + V 2 ) 2 ,
V i =   ( V 2 V 1 ) 2 2 ,
where V 1 ( m m 3 ) and V 2 ( m m 3 ) denote the volumes from the same tibia in scan 1 and scan 2, respectively, which are, thus, measured twice on different days yielding a single paired observation. The square of the within-subject coefficient of variation ( wCV 2 ) is first obtained by taking the mean of the ratio of variance V i to the square of the mean M i 2 over all N pairs and then is applied to calculate the within-subject coefficient of variation (wCV%) as a percentage as follows:
w C V 2 = 1 N i N V i M i 2 ,
w C V % = 100 × w C V 2 .
The corresponding 95% confidence intervals (α = 0.05) to wCV are given by the multiplicative lower-bound (LB) and upper-bound (UB) factors given by ChiSqr function for N-1 degrees of freedom as
95 %   CI   of   wCV =   wCV % · [ 1 ( N 1 ) ·   χ 0.975 2 ;   1 ( N 1 ) ·   χ 0.025 2 ] .
Bland–Altman analysis (BA):
The BA agreement was computed for the pairwise volume differences ( V 2 V 1 ) i against their means M i . The mean of all N paired differences on tibia volume data is assessed to measure the bias between test and retest volumes. Ideally, the bias should be close to zero thereby supporting the assumption that the volume will remain constant between test and retest measurements.
B i a s = 1 N i N   ( V 2 V 1 ) i .
The corresponding 95% limits of agreement (LOA) along the difference axis ( V 2 V 1 ) in Bland–Altman plots provide a graphical indication of measurement precision. To find the precision, the standard deviation SD of the differences between test and retest volumes is obtained, and then the 95% LOA of ( V 2 V 1 ) is calculated as follows:
P r e c i s i o n   ( L O A ) = B i a s   ± 1.96 S D   ( V 2 V 1 ) .

References

  1. Gangat, N.; Caramazza, D.; Vaidya, R.; George, G.; Begna, K.; Schwager, S.; Van Dyke, D.; Hanson, C.; Wu, W.; Pardanani, A.; et al. DIPSS Plus: A Refined Dynamic International Prognostic Scoring System for Primary Myelofibrosis That Incorporates Prognostic Information from Karyotype, Platelet Count, and Transfusion Status. J. Clin. Oncol. 2011, 29, 392–397. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, F.; Qiu, T.; Wang, H.; Yang, Q. State-of-the-Art Review on Myelofibrosis Therapies. Clin. Lymphoma Myeloma Leuk. 2022, 22, e350–e362. [Google Scholar] [CrossRef] [PubMed]
  3. Gianelli, U.; Vener, C.; Bossi, A.; Cortinovis, I.; Iurlo, A.; Fracchiolla, N.S.; Savi, F.; Moro, A.; Grifoni, F.; De Philippis, C.; et al. The European Consensus on grading of bone marrow fibrosis allows a better prognostication of patients with primary myelofibrosis. Mod. Pathol. 2012, 25, 1193–1202. [Google Scholar] [CrossRef] [Green Version]
  4. Leimkühler, N.B.; Gleitz, H.F.; Ronghui, L.; Snoeren, I.A.; Fuchs, S.N.; Nagai, J.S.; Banjanin, B.; Lam, K.H.; Vogl, T.; Kuppe, C.; et al. Heterogeneous bone-marrow stromal progenitors drive myelofibrosis via a druggable alarmin axis. Cell Stem Cell 2021, 28, 637–652.e8. [Google Scholar] [CrossRef]
  5. Luker, G.D.; Nguyen, H.; Hoff, B.A.; Galbán, C.J.; Hernando, D.; Chenevert, T.L.; Talpaz, M.; Ross, B.D. A Pilot Study of Quantitative MRI Parametric Response Mapping of Bone Marrow Fat for Treatment Assessment in Myelofibrosis. Tomography 2016, 2, 67–78. [Google Scholar] [CrossRef] [PubMed]
  6. Robison, T.H.; Solipuram, M.; Heist, K.; Amouzandeh, G.; Lee, W.Y.; Humphries, B.A.; Buschhaus, J.M.; Bevoor, A.; Zhang, A.; Luker, K.E.; et al. Multiparametric MRI to quantify disease and treatment response in mice with myeloproliferative neoplasms. JCI Insight 2022, 7, 161457. [Google Scholar] [CrossRef]
  7. Ross, B.D.; Jang, Y.; Welton, A.; Bonham, C.A.; Palagama, D.S.W.; Heist, K.; Boppisetti, J.; Imaduwage, K.P.; Robison, T.; King, L.R.; et al. A lymphatic-absorbed multi-targeted kinase inhibitor for myelofibrosis therapy. Nat. Commun. 2022, 13, 4730. [Google Scholar] [CrossRef] [PubMed]
  8. Obuchowski, N.A. Interpreting Change in Quantitative Imaging Biomarkers. Acad. Radiol. 2018, 25, 372–379. [Google Scholar] [CrossRef] [PubMed]
  9. Raunig, D.L.; McShane, L.M.; Pennello, G.; Gatsonis, C.; Carson, P.L.; Voyvodic, J.T.; Wahl, R.L.; Kurland, B.F.; Schwarz, A.J.; Gönen, M.; et al. Quantitative imaging biomarkers: A review of statistical methods for technical performance assessment. Stat. Methods Med. Res. 2015, 24, 27–67. [Google Scholar] [CrossRef] [Green Version]
  10. Jha, A.K.; Mithun, S.; Jaiswar, V.; Sherkhane, U.B.; Purandare, N.C.; Prabhash, K.; Rangarajan, V.; Dekker, A.; Wee, L.; Traverso, A. Repeatability and reproducibility study of radiomic features on a phantom and human cohort. Sci. Rep. 2021, 11, 2055. [Google Scholar] [CrossRef]
  11. Malaih, A.A.; Dunn, J.T.; Nygård, L.; Kovacs, D.G.; Andersen, F.L.; Barrington, S.F.; Fischer, B.M. Test–retest repeatability and interobserver variation of healthy tissue metabolism using 18F-FDG PET/CT of the thorax among lung cancer patients. Nucl. Med. Commun. 2022, 43, 549–559. [Google Scholar] [CrossRef] [PubMed]
  12. Shukla-Dave, A.; Obuchowski, N.A.; Chenevert, T.L.; Jambawalikar, S.; Schwartz, L.H.; Malyarenko, D.; Huang, W.; Noworolski, S.M.; Young, R.J.; Shiroishi, M.S.; et al. Quantitative imaging biomarkers alliance (QIBA) recommendations for improved precision of DWI and DCE-MRI derived biomarkers in multicenter oncology trials. J. Magn. Reson. Imaging 2019, 49, e101–e121. [Google Scholar] [CrossRef] [PubMed]
  13. Newitt, D.C.; Amouzandeh, G.; Partridge, S.C.; Marques, H.S.; Herman, B.A.; Ross, B.D.; Hylton, N.M.; Chenevert, T.L.; Malyarenko, D.I. Repeatability and Reproducibility of ADC Histogram Metrics from the ACRIN 6698 Breast Cancer Therapy Response Trial. Tomography 2020, 6, 177–185. [Google Scholar] [CrossRef] [PubMed]
  14. Amouzandeh, G.; Heist, K.A.; Malyarenko, D.I.; Jang, J.; Robison, T.; Bonham, C.; Amirfazli, C.; Swanson, S.D.; Luker, G.D.; Ross, B.D.; et al. MR Imaging of Murine Tibia for Co-Clinical Studies of Myelofibrosis. In Proceedings of the International Society of Magnetic Resonance in Medicine Annual Meeting & Exhibition, Co-Clinical Imaging Research Symposium, Virtual, 15–20 May 2021; p. 2355. [Google Scholar]
  15. Becker, A.S.; Chaitanya, K.; Schawkat, K.; Muehlematter, U.J.; Hötker, A.M.; Konukoglu, E.; Donati, O.F. Variability of manual segmentation of the prostate in axial T2-weighted MRI: A multi-reader study. Eur. J. Radiol. 2019, 121, 108716. [Google Scholar] [CrossRef]
  16. Newitt, D.C.; Zhang, Z.; Ba, J.E.G.; Partridge, S.C.; Chenevert, T.L.; Rosen, M.A.; Bolan, P.; Marques, H.S.; Aliu, S.; Li, W.; et al. Test–retest repeatability and reproducibility of ADC measures by breast DWI: Results from the ACRIN 6698 trial. J. Magn. Reson. Imaging 2019, 49, 1617–1628. [Google Scholar] [CrossRef] [PubMed]
  17. Qiu, B.; van der Wel, H.; Kraeima, J.; Glas, H.H.; Guo, J.; Borra, R.J.H.; Witjes, M.J.H.; van Ooijen, P.M.A. Automatic Segmentation of Mandible from Conventional Methods to Deep Learning—A Review. J. Pers. Med. 2021, 11, 629. [Google Scholar] [CrossRef] [PubMed]
  18. Renard, F.; Guedria, S.; De Palma, N.; Vuillerme, N. Variability and reproducibility in deep learning for medical image segmentation. Sci. Rep. 2020, 10, 13724. [Google Scholar] [CrossRef]
  19. Ma, X.; Hadjiiski, L.M.; Wei, J.; Chan, H.; Cha, K.H.; Cohan, R.H.; Caoili, E.M.; Samala, R.; Zhou, C.; Lu, Y. U-Net based deep learning bladder segmentation in CT urography. Med. Phys. 2019, 46, 1752–1765. [Google Scholar] [CrossRef]
  20. Wang, Y.; Zhou, C.; Chan, H.-P.; Hadjiiski, L.M.; Chughtai, A. Fusion of multiple deep convolutional neural networks (DCNNs) for improved segmentation of lung nodules in CT images. In Proceedings of the SPIE Medical Imaging 2022, San Diego, CA, USA, 20–24 February 2022; Volume 12033. [Google Scholar] [CrossRef]
  21. Chan, H.-P.; Samala, R.K.; Hadjiiski, L.M.; Zhou, C. Deep Learning in Medical Image Analysis. Adv. Exp. Med. Biol. 2020, 1213, 3–21. [Google Scholar] [CrossRef]
  22. Heckelman, L.N.; Soher, B.J.; Spritzer, C.E.; Lewis, B.D.; DeFrate, L.E. Design and validation of a semi-automatic bone segmentation algorithm from MRI to improve research efficiency. Sci. Rep. 2022, 12, 7825. [Google Scholar] [CrossRef]
  23. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef] [PubMed]
  24. Kushwaha, A.; Mourad, R.F.; Heist, K.; Malyarenko, D.; Chan, H.P.; Chenevert, T.L.; Hadjiiski, L.M. Segmentation of Mouse Tibia on MRI using Deep Learning U-Net Models. In Proceedings of the SPIE Medical Imaging, San Deigo, CA, USA, 19–23 February 2023. [Google Scholar]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention Part III, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  26. Obuchowski, N.A.; Bullen, J. Quantitative imaging biomarkers: Effect of sample size and bias on confidence interval coverage. Stat. Methods Med. Res. 2018, 27, 3139–3150. [Google Scholar] [CrossRef] [PubMed]
  27. Mali, S.A.; Ibrahim, A.; Woodruff, H.C.; Andrearczyk, V.; Müller, H.; Primakov, S.; Salahuddin, Z.; Chatterjee, A.; Lambin, P. Making Radiomics More Reproducible across Scanner and Imaging Protocol Variations: A Review of Harmonization Methods. J. Pers. Med. 2021, 11, 842. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, S.; Li, C.; Wang, R.; Liu, Z.; Wang, M.; Tan, H.; Wu, Y.; Liu, X.; Sun, H.; Yang, R.; et al. Annotation-efficient deep learning for automatic medical image segmentation. Nat. Commun. 2021, 12, 5915. [Google Scholar] [CrossRef]
  29. Candemir, S.; Nguyen, X.V.; Folio, L.R.; Prevedello, L.M. Training Strategies for Radiology Deep Learning Models in Data-limited Scenarios. Radiol. Artif. Intell. 2021, 3, e210014. [Google Scholar] [CrossRef]
  30. Madireddy, I.; Wu, T. Rule and Neural Network-Based Image Segmentation of Mice Vertebrae Images. Cureus 2022, 14, e27247. [Google Scholar] [CrossRef]
  31. Hadjiiski, L.; Cha, K.; Chan, H.; Drukker, K.; Morra, L.; Näppi, J.J.; Sahiner, B.; Yoshida, H.; Chen, Q.; Deserno, T.M.; et al. AAPM task group report 273: Recommendations on best practices for AI and machine learning for computer-aided diagnosis in medical imaging. Med. Phys. 2023, 50, e1–e24. [Google Scholar] [CrossRef]
  32. Holbrook, M.D.; Blocker, S.J.; Mowery, Y.M.; Badea, A.; Qi, Y.; Xu, E.S.; Kirsch, D.G.; Johnson, G.A.; Badea, C. MRI-Based Deep Learning Segmentation and Radiomics of Sarcoma in Mice. Tomography 2020, 6, 23–33. [Google Scholar] [CrossRef]
  33. Seo, S.Y.; Kim, S.-J.; Oh, J.S.; Chung, J.; Kim, S.-Y.; Oh, S.J.; Joo, S.; Kim, J.S. Unified Deep Learning-Based Mouse Brain MR Segmentation: Template-Based Individual Brain Positron Emission Tomography Volumes-of-Interest Generation Without Spatial Normalization in Mouse Alzheimer Model. Front. Aging Neurosci. 2022, 14, 807903. [Google Scholar] [CrossRef]
  34. Kenney, H.M.; Peng, Y.; Chen, K.L.; Ajalik, R.; Schnur, L.; Wood, R.W.; Schwarz, E.M.; Awad, H.A. A high-throughput semi-automated bone segmentation workflow for murine hindpaw micro-CT datasets. Bone Rep. 2022, 16, 101167. [Google Scholar] [CrossRef]
  35. Schoppe, O.; Pan, C.; Coronel, J.; Mai, H.; Rong, Z.; Todorov, M.I.; Müskes, A.; Navarro, F.; Li, H.; Ertürk, A.; et al. Deep learning-enabled multi-organ segmentation in whole-body mouse scans. Nat. Commun. 2020, 11, 5626. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Experimental design diagram showing data content (including repeatability scan pairs) for different A-U-net “training” split scenarios (lower left branches). Data subsets that had manual segmentations by Expert Annotator 1 and 2 (EA1 and EA2) available are marked in the middle branches. The inset shows an example of 3D MRI image for mouse tibia with three orthogonal view projection planes.
Figure 1. Experimental design diagram showing data content (including repeatability scan pairs) for different A-U-net “training” split scenarios (lower left branches). Data subsets that had manual segmentations by Expert Annotator 1 and 2 (EA1 and EA2) available are marked in the middle branches. The inset shows an example of 3D MRI image for mouse tibia with three orthogonal view projection planes.
Tomography 09 00048 g001
Figure 2. (A) Attention U-net architecture with (B) attention gate diagram: the black arrows indicate the workflow direction; yellow boxes depict the operation output (with the image dimensions (height × width) on the side and the number of applied filters shown on top), and green boxes correspond to output from the attention gate (B); colored triangles denote operations described in the legends. Output from the attention gate (green box) and up convolutions (yellow triangle) are concatenated (merged boxes). Attention coefficients are 2D arrays obtained from the input 2D MRI slices by the set of operations shown in (B).
Figure 2. (A) Attention U-net architecture with (B) attention gate diagram: the black arrows indicate the workflow direction; yellow boxes depict the operation output (with the image dimensions (height × width) on the side and the number of applied filters shown on top), and green boxes correspond to output from the attention gate (B); colored triangles denote operations described in the legends. Output from the attention gate (green box) and up convolutions (yellow triangle) are concatenated (merged boxes). Attention coefficients are 2D arrays obtained from the input 2D MRI slices by the set of operations shown in (B).
Tomography 09 00048 g002
Figure 3. Examples of mouse tibia segmentation contours (magenta) for a single-slice MRI image in the test set (A) delineated by expert annotators ((B) EA1; (F): EA2) and by DL model with different training scenarios; (C) full training subset; (D) TS1; (G) TSM; (H) TS2). (E) The full tibia 3D volume segmentation by EA1 for two scans (6 weeks apart) of the same mouse (top = 6.5 mm3; lower = 9.9 mm3).
Figure 3. Examples of mouse tibia segmentation contours (magenta) for a single-slice MRI image in the test set (A) delineated by expert annotators ((B) EA1; (F): EA2) and by DL model with different training scenarios; (C) full training subset; (D) TS1; (G) TSM; (H) TS2). (E) The full tibia 3D volume segmentation by EA1 for two scans (6 weeks apart) of the same mouse (top = 6.5 mm3; lower = 9.9 mm3).
Tomography 09 00048 g003
Figure 4. Examples of mouse tibia segmentation contours (magenta) for a single-slice MRI image in the test set (A) delineated by expert annotators ((B) EA1; (F) EA2) and by DL model with different training scenarios (C) full training; (D) TS1; (G) TSM; (H) TS2). (E) The full tibia 3D volume (6.5 mm3) segmentation by EA1. The yellow arrows indicate apparent tibia under-segmentations in (B) and (G).
Figure 4. Examples of mouse tibia segmentation contours (magenta) for a single-slice MRI image in the test set (A) delineated by expert annotators ((B) EA1; (F) EA2) and by DL model with different training scenarios (C) full training; (D) TS1; (G) TSM; (H) TS2). (E) The full tibia 3D volume (6.5 mm3) segmentation by EA1. The yellow arrows indicate apparent tibia under-segmentations in (B) and (G).
Tomography 09 00048 g004
Figure 5. Tibia volume (including test–retest) accuracy for the test subset with respect to EA1 reference for different training scenarios (color-coded in the legend). Solid horizontal lines mark the bias (average volume difference), and dashed lines correspond to 95% limits of agreement.
Figure 5. Tibia volume (including test–retest) accuracy for the test subset with respect to EA1 reference for different training scenarios (color-coded in the legend). Solid horizontal lines mark the bias (average volume difference), and dashed lines correspond to 95% limits of agreement.
Tomography 09 00048 g005
Figure 6. Bland–Altman (BA) analysis of the scan–rescan (test–retest) tibia volumes in the test dataset for different A-U-net training scenarios ((A) full training set; (D) half-split 1; (E) half-split 2; (F) single-mouse) and human expert annotations ((B) EA1; (C) EA2) segmentations. Solid horizontal lines mark the observed average bias of the volume difference, while dashed lines mark corresponding 95% limits of agreement.
Figure 6. Bland–Altman (BA) analysis of the scan–rescan (test–retest) tibia volumes in the test dataset for different A-U-net training scenarios ((A) full training set; (D) half-split 1; (E) half-split 2; (F) single-mouse) and human expert annotations ((B) EA1; (C) EA2) segmentations. Solid horizontal lines mark the observed average bias of the volume difference, while dashed lines mark corresponding 95% limits of agreement.
Tomography 09 00048 g006
Table 1. Summary of segmentation accuracy metrics (mean ± standard deviation (SD)) for both experts and DL model with different training split scenarios (full, split 1, split 2, and single mouse).
Table 1. Summary of segmentation accuracy metrics (mean ± standard deviation (SD)) for both experts and DL model with different training split scenarios (full, split 1, split 2, and single mouse).
DatasetAJI%AVI %AVE%AHD (mm)
Attention U-Net trained on full training setTraining88.63 ± 1.2194.64 ± 1.64−1.44 ± 3.070.20 ± 0.05
Validation83.08 ± 2.6891.06 ± 3.79−0.67 ± 6.280.52 ± 0.30
Test83.45 ± 5.1190.21 ± 6.281.71 ± 7.940.47 ± 0.43
Attention U-Net trained on split1Training86.17 ± 2.2293.51 ± 2.74−2.06 ± 4.920.39 ± 0.68
Validation81.47 ± 4.5889.27 ± 6.181.22 ± 8.650.57 ± 0.28
Test82.47 ± 6.7989.15 ± 8.172.83 ± 9.890.79 ± 1.39
Attention U-Net trained on split2Training88.34 ± 1.1594.79 ± 1.57−2.11 ± 3.190.23 ± 0.06
Validation82.78 ± 2.6591.15 ± 3.85−1.29 ± 6.990.78 ± 0.77
Test83.27 ± 5.0189.94 ± 6.392.11 ± 8.350.49 ± 0.42
Attention U-Net trained on single mouseTraining86.38 ± 0.6997.32 ± 1.09−9.99 ± 2.970.29 ± 0.22
Validation76.42 ± 6.4987.70 ± 8.45−2.41 ± 11.631.45 ± 1.41
Test77.99 ± 7.4387.77 ± 10.05−0.15 ± 14.311.10 ± 1.07
EA1 vs. EA2 (EA2 reference)TrainingNA
Validation77.98 ± 2.6397.39 ± 1.39−22.44 ± 6.460.36 ± 0.15
Test80.70 ± 2.9196.70 ± 2.33−16.70 ± 7.660.29 ± 0.11
EA2 vs. EA1 (EA1 reference)TrainingNA
Validation77.98 ± 2.6479.70 ± 3.3618.11 ± 4.390.36 ± 0.15
Test80.70 ± 2.9183.10 ± 3.9213.94 ± 5.760.29 ± 0.11
AJI: average Jaccard index, AVI: average volume intersection ratio, AVE: average volume error, AHD: average Hausdorff distance.
Table 2. The repeatability of volume estimations in terms of %wCV and confidence intervals [CIs] in all data subsets for A-U-net model and EA1, EA2 expert annotations and different DL model training scenarios (Figure 1: full, TS1, TS2, TSM).
Table 2. The repeatability of volume estimations in terms of %wCV and confidence intervals [CIs] in all data subsets for A-U-net model and EA1, EA2 expert annotations and different DL model training scenarios (Figure 1: full, TS1, TS2, TSM).
ModelTrainingValidationTest
wCV [CI] %
A-U-netFull7.7 [6.2, 10.3]
7.6 [5.5, 12.0]
8.0 [6.0, 12.3]
3.3 [2, 12]
2.0 [1.1, 7.4]
5.3 [3.0, 20.0]
1.83 [1.0, 6.8]
11.1 [6.3, 41.6]
2.6 [1.9, 4.3]
3.1 [2.2, 5.1]
3.2 [2.3, 5.3]
7.0 [5.0, 11.6]
TS1
TS2
TSM
EA1Full7.7 [6.2, 10.3]
7.3 [5.4, 11.5]
8.0 [6.0, 12.3]
5.8 [4.6, 7.7]
4.4 [2.5, 16.3]5.3 [3.8, 8.7]
TS1
TS2
TSM
EA2FullNA2.5 [1.4, 9.1]8.0 [5.7, 13.2]
TS1
TS2
TSM
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kushwaha, A.; Mourad, R.F.; Heist, K.; Tariq, H.; Chan, H.-P.; Ross, B.D.; Chenevert, T.L.; Malyarenko, D.; Hadjiiski, L.M. Improved Repeatability of Mouse Tibia Volume Segmentation in Murine Myelofibrosis Model Using Deep Learning. Tomography 2023, 9, 589-602. https://doi.org/10.3390/tomography9020048

AMA Style

Kushwaha A, Mourad RF, Heist K, Tariq H, Chan H-P, Ross BD, Chenevert TL, Malyarenko D, Hadjiiski LM. Improved Repeatability of Mouse Tibia Volume Segmentation in Murine Myelofibrosis Model Using Deep Learning. Tomography. 2023; 9(2):589-602. https://doi.org/10.3390/tomography9020048

Chicago/Turabian Style

Kushwaha, Aman, Rami F. Mourad, Kevin Heist, Humera Tariq, Heang-Ping Chan, Brian D. Ross, Thomas L. Chenevert, Dariya Malyarenko, and Lubomir M. Hadjiiski. 2023. "Improved Repeatability of Mouse Tibia Volume Segmentation in Murine Myelofibrosis Model Using Deep Learning" Tomography 9, no. 2: 589-602. https://doi.org/10.3390/tomography9020048

APA Style

Kushwaha, A., Mourad, R. F., Heist, K., Tariq, H., Chan, H. -P., Ross, B. D., Chenevert, T. L., Malyarenko, D., & Hadjiiski, L. M. (2023). Improved Repeatability of Mouse Tibia Volume Segmentation in Murine Myelofibrosis Model Using Deep Learning. Tomography, 9(2), 589-602. https://doi.org/10.3390/tomography9020048

Article Metrics

Back to TopTop