Next Article in Journal
A Meta-Analysis of Computerized Tomography-Based Radiomics for the Diagnosis of COVID-19 and Viral Pneumonia
Next Article in Special Issue
Laboratory Demand Management Strategies—An Overview
Previous Article in Journal
Cardiac Manifestations in a Group of Romanian Patients with Gaucher Disease Type 1 (a Monocentric Study)
Previous Article in Special Issue
Has the Flood Entered the Basement? A Systematic Literature Review about Machine Learning in Laboratory Medicine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep into Laboratory: An Artificial Intelligence Approach to Recommend Laboratory Tests

by
Md. Mohaimenul Islam
1,2,3,
Tahmina Nasrin Poly
1,2,3,
Hsuan-Chia Yang
1,2,3 and
Yu-Chuan (Jack) Li
1,2,3,4,5,*
1
Graduate Institute of Biomedical Informatics, College of Medical Science and Technology, Taipei Medical University, Taipei 110, Taiwan
2
International Center for Health Information Technology (ICHIT), Taipei Medical University, Taipei 110, Taiwan
3
Research Center of Big Data and Meta-Analysis, Wan Fang Hospital, Taipei Medical University, Taipei 116, Taiwan
4
Department of Dermatology, Wan Fang Hospital, Taipei 116, Taiwan
5
TMU Research Center of Cancer Translational Medicine, Taipei Medical University, Taipei 110, Taiwan
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(6), 990; https://doi.org/10.3390/diagnostics11060990
Submission received: 29 March 2021 / Revised: 19 May 2021 / Accepted: 27 May 2021 / Published: 29 May 2021

Abstract

:
Laboratory tests are performed to make effective clinical decisions. However, inappropriate laboratory test ordering hampers patient care and increases financial burden for healthcare. An automated laboratory test recommendation system can provide rapid and appropriate test selection, potentially improving the workflow to help physicians spend more time treating patients. The main objective of this study was to develop a deep learning-based automated system to recommend appropriate laboratory tests. A retrospective data collection was performed at the National Health Insurance database between 1 January 2013, and 31 December 2013. We included all prescriptions that had at least one laboratory test. A total of 1,463,837 prescriptions from 530,050 unique patients was included in our study. Of these patients, 296,541 were women (55.95%), the range of age was between 1 and 107 years. The deep learning (DL) model achieved a higher area under the receiver operating characteristics curve (AUROC micro = 0.98, and AUROC macro = 0.94). The findings of this study show that the DL model can accurately and efficiently identify laboratory tests. This model can be integrated into existing workflows to reduce under- and over-utilization problems.

1. Introduction

The laboratory test plays a vital role in patient care, accurate diagnosis, and treatment planning (approximately 60–70% of the clinical decision is made based on laboratory tests) [1,2]. Previous studies report that the rate of laboratory test ordering has been increased in general practice [3,4,5]; although they are frequently overused, misused, and even underused, which raises questions [6,7]. A higher proportion of laboratory test errors occur in the pre-analytical stage, leading to misdirected or delayed diagnosis, jeopardizing patients’ safety. An inappropriate laboratory test is not only associated with an increased risk of psychological and physical harm [8,9] but also causes unnecessary financial burden [10,11]. Unfortunately, there is no direct relationship between higher spending and healthcare benefits.
Inadequate knowledge of laboratory tests, availability of higher numbers of new tests, fear of diagnostic errors, and lack of education have been found to be associated with inappropriate laboratory testing [12,13]. Several factors also increased the rate of inappropriate laboratory testing, such as increasing patient satisfaction, fear of liability, and fee-for-service billing [10,14]. Existing strategies to reduce inappropriate tests and improve the efficiency of ordering clinical laboratory tests include feedback and reminder [15,16], clinical decision support tools [17], education [18], cost display [19], practice guidelines [20], and condition-specific algorithms [21]. The performance of these interventions varies, depending on how these interventions are applied.
Nowadays, the demand for automated laboratory recommendation systems for accurate and quicker diagnosis has been increased. An automated recommendation might improve the accuracy and efficiency of test orders and contribute to healthcare resource-saving. In this study, we developed a novel automated recommendation system-based deep neural network (DNN) that automatically predicts accurate and efficient laboratory tests using simple variables available in the electronic health records (EHRs). We evaluated DNN model performance and assessed the acceptability, generalizability, and interpretability by using standard guidelines.

2. Methods

2.1. Data Source

We collected data from the Taiwanese National Health Insurance Research and Development (NHIRD). The NHIRD stores all claims of medications, diagnoses, and laboratory work for 23 million people in Taiwan, which represents approximately 99.9% of the total population [22]. The NHIRD database includes patients’ demographic information, total number of prescriptions, date of prescriptions, number of medications, dosages, durations, and comorbidities history. The quality and completeness of the NHIRD database is excellent and it is used to conduct high-quality clinical researches [23,24]. The findings from these researches have been used in valuable clinical decision-making, patient care improvement, and policy-making. The Taipei Medical University (TMU) research ethical board approved this study. Participants’ consent was not required because the entire individual’s information was de-identified.

2.2. Data Collection and Study Populations

In this study, we collected patients’ demographics, medications, diseases, and laboratory information from a random sample of two million from the NHIRD between 1 January 2013, and 31 December 2013. We selected all prescriptions that contained at least one laboratory test.

2.3. Data Preprocessing

Data preprocessing consists of three steps: (1) data cleaning, (2) variable selection, and (3) matrix formation. NHIRD stores a vast amount of variables but all variables are not necessary for doing research. We first deleted unnecessary variables and kept only demographics, visit identification, visit date, date of birth, department identification, medications, diseases, and laboratory information. Then we selected patients who ordered at least one laboratory test on each visit. Afterwards, we calculated an age for all patients using their birth date and visit date information. There was no age limitation included in our work. The International Classification of Diseases, Ninth Revision, Clinical Modification (ICD-9-CM) code consists of a five-digit code; the first three digits describe the category, and the last two digits describe aetiology, anatomic site, and manifestation (Figure 1). For example, ICD-9 of diabetes mellitus is 250. However, it has further been classified into several groups, such as 250.0: Diabetes mellitus without mention of complication, type II or unspecified type, not stated as uncontrolled, 250.1: Diabetes with ketoacidosis, and 250.10: Diabetes with ketoacidosis, type II or unspecified type, not stated as uncontrolled.
However, we selected three digits of ICD-9-CM, including the V-code (supplementary classification of factors influencing health status and contact with health services) classification to retrieved comorbidities information. They were usually distributed from 001~999 and V01~V82. In the case of Anatomical Therapeutic Chemical Classification (ATC), five characters of ATC code were considered in this study (Figure 2), e.g., five-digit ATC code: A10BA (Biguanides), blood glucose lowering drugs, including all drugs like A10BA01 (phenformin), A10BA02 (metformin), and A10BA03 (buformin). Also, seven characters (e.g., A10BA02) were considered as chemical substances even though five characters were used to describe chemical substances and all drugs included in this group were prescribed for almost the same purpose. There were 654 different types of laboratory tests ordered during the study period. We calculated the frequency of laboratory tests and selected those tests that contributed at least 150 prescriptions (0.001 per cent) during our study period. This is because a deep learning algorithm needs sufficient data to train the model, otherwise, it can perform poorly. Finally, 315 types of tests were included in our study.
In the multi-label classification, each instance belongs to several labels simultaneously (Figure 3). In our case, each laboratory test represents a label, and a patient can have different types of laboratory test at the same time. For example, if the data contains four labels, such as glucose, HbA1c, urine, and creatinine, then that patient can have 24 types of outcomes, such as only glucose, only HbA1c, only urine, only creatinine, or all glucose, HbA1c, urine, and creatinine, or glucose and HbA1c, and so forth and so on.

2.4. Model Building

We split data set into 70% training set and 30% testing set. In the internal validation, 20% of the data was randomly selected from the training set and evaluated against model performance (Figure 4).
The DNN model was developed to train all variables and the model was assessed using the validation set to predict laboratory tests. DNN is a high-performing algorithm in which an artificial neural network consists of multiple layers. The input of DNN moves through the layers calculating the probability of each output. We used three hidden layers in our model. The activation function ReLU was used in the hidden layers and Sigmoid was used in the output layer. The activation function is an integral part of a neural network that is often known for non-linearity, i.e., describing the input and output relations in a non-linear way. However, the non-linearity element allows for higher flexibility and makes a complex function during the whole model learning process. It helps to speed up the whole learning process. Several activation functions, such as sigmoid and ReLU, are commonly used in practice.
(a) Sigmoid Function:
This takes a real-value input and converts it to a range between 0 and 1. The sigmoid function is defined as follows:
σ ( x ) = 1 1 + e x
Here it is clear that it will convert output between 0 and 1 when the input varies in ( , ) . A neuron can use the sigmoid for computing the nonlinear function σ   ( y = w x + b ) . If y = w x + b is very large and positive, then e y 0 , so σ ( y ) 1 , while y = w x + b is very large and negative e y , so σ ( y ) 0 .
(b) ReLU: This stands for Rectified Linear Unit and it takes a real input variable and thresholds it at zero (replacing native values with zero). The ReLU function is defined as follows:
f = max ( 0 ,   x )
A total of 1000 epochs was utilized to calculate training and validation loss (Figure 5).

2.5. Evaluation Matrices

We evaluated the performance of our model on the validations set for laboratory tests with recommendations using the following metrics.
Micro-AUC: This is an averaging of the prediction matrix. Smicro is the set of correct quadruples. The equation of assessing Micro-AUC is as follows:
M i c r o - A U C = | S m i c r o | ( i = 1 m | Y i . + | ) · ( i = 1 m |   Y i . | )
S m i c r o = { ( a , b , i , j ) | ( a , b ) Y . i + × Y . j ,   f i ( x a ) f j ( x b ) }
Macro-AUC = This is an averaging of each label. S m a c r o is the set of correctly ordered instance pairs on each label. The equation for assessing macro-AUC is as follows:
M a c r o - A U C = 1 l   j = 1 l | S m a c r o j | | Y . j + | | Y . J |
S m a c r o j = { ( a , b ) Y . j + × Y . j | f i ( x a ) f i ( x b ) }
Micro-F1: This is an averaging of the prediction matrix. The equation is as follows:
M i c r o - F 1 = 2 j = 1 l i = 1 m y i j h i j j = 1 l i = 1 l y i j + j = 1 l i = 1 m h i j
Macro-F1: This is an averaging of each label. The equation for assessing macro-F1 is as follows:
M a c r o - F 1 = 1 l   j = 1 l 2 i = 1 m y i j h i j i = 1 m y i j + i = 1 m h i j
Average Precision: Average precision summarizes the fraction of relevant labels ranked higher than one other relevant label. The equation is as follows:
A v e r a g e   P r e c i s i o n = 1 m i = 1 m 1 | Y i . + |   j Y i . + | S p r e c i s i o n i , j | r a n k F ( x i   ,   j )
S p r e c i s i o n i j = { k Y i . + | r a n k F ( x i   , k ) r a n k F ( x i ,   j ) }
Hamming Loss (HL): This is the average symmetric difference between the set of true labels and the set of predicted labels of the data set. The equation is as follows:
H l o s s   ( H ) = 1 m l   i = 1 m j 1 l h i j y i j
HL value ranges from 0 to 1. A lesser value of hamming loss indicates a better classifier. The analyses were performed using Python (version 3.7.3; https://www.python.org (accessed on 1 January 2020)).

3. Results

3.1. Data Descriptions

We collected all patients’ data retrospectively who visited a hospital between 1 January 2013 and 31 December 2013. A total of 536,015 (26.80%) patients were ordered laboratory tests, consisting of 1,491,564 prescriptions. After excluding infrequent laboratory test orders, we finally considered 315 different types of laboratory tests from 530,050 patients with 1,463,837 prescriptions to develop our prediction model. The number of female patients was higher than male patients (44.05% [233,509/530,050] vs. 55.95% [296,541/530,050] and the age of patients ranged from 1 to 107 years.

3.2. Model Performance

The DL model demonstrated the highest discriminating power for all diagnostic laboratory tests using minimal variables from EHR, yielding a mean AUROC micro of 0.98 and AUROC macro of 0.94, respectively (Figure 6).
Table 1 presents the performance of the DL model for the prediction of laboratory tests in varying cut-off points. Our prediction model showed the highest recall 0.96 when the cut-off value 0.01 was considered. However, precision, recall, and F1 score value were almost similar at the cut-off value 0.30 and 0.35. Hamming loss significantly was going down while increasing cut-off value reached 0.011 when cut-off value 0.4 was considered.

3.3. Distribution of AUROC

The DL model achieved more than 80% AUROC for the prediction of every laboratory test except two tests named urine biochemistry examination and CBC-Ⅲ. The range of AUROC was between 0.76 and 1. However, the DL model achieved AUROC of 0.92–0.96 for 114 laboratory tests, 0.96–1 for 106 laboratory tests, and 0.88–0.92, 0.84–0.88, 0.80–0.84, and 0.76–0.80 for 56, 30, five, and four laboratory tests, respectively (Figure 7).

3.4. Evaluation

After developing the DL model to predict laboratory tests, we evaluated the effectiveness of our model. We used the “lab test online” for evaluating laboratory testing appropriateness, which contains all information, including disease and various laboratory tests in detail. We selected some random cases from the testing set for preliminary evaluations to evaluate our model performance. Figure 8 shows one example of evaluation of the DL model. Here, an 80-year-old male patient had four diseases: iron deficiency anemia, acute myocardial infarction, cystitis, and hyperplasia of the prostate. The physician prescribed several medications for his problems and ordered 12 laboratory tests based on the patient’s physical condition. Our model predicted the same laboratory tests while considering cut-off value 0.35 and 0.30. However, our model recommends two additional laboratory tests while considering cut-off value 0.30 for a second patient. There are credible reasons for these two tests in this situation because a patient with rheumatoid arthritis and other inflammatory polyarthropathies needs to check WBC differential count on a regular basis. Rheumatoid factor test–Nephelometry is often used when a patient is suspected of having a kidney problem. The physician may adjust cut-off value when more tests are needed because the laboratory test depends on the patient’s specific condition.

4. Discussion

4.1. Main Findings

Our primary objective was to reduce inappropriate laboratory test ordering by developing a DL model based on simple variables available in the EHR. This study represents the first investigation using the DL model to develop and internally validate a diagnostic laboratory tests prediction model. The DL model has great potential for predicting events and recognizing data patterns used in many healthcare industry areas. Our findings are fascinating since the DL model showed an excellent caliber for predicting laboratory tests (AUROC micro = 0.98 and AUROC macro = 0.94). The range of AUROC was between 0.76 and 1. This study suggests considering a threshold between 0.30 and 0.35 to achieve optimal laboratory test recommendations. However, the recall is higher (more than 0.95) when the threshold is 0.01, which would recommend all possible laboratory tests in the recommendation system. All these laboratory tests have credible reasons for the recommendation, and there is no issue for false negatives.
Defining appropriate laboratory test orders is not an easy task and is influenced by several factors, but the findings of our study show that the DL model can predict appropriate laboratory tests and ensure personalized treatment. Only four features, namely patients’ age, gender, diseases, and drugs history, are synthesized through the DL model to make vigorous and accurate predictions. Various applications and clinical contexts might need a different level of acceptability, therefore, this study gives all possible recommendations while choosing a potential target recall close to 96% (when our model predicts laboratory tests result as correct, the goal was for it to be right 96% of the time). Moreover, the results of our study suggest that our model may be implemented in real-world clinical settings because it had only a 4% false-negative prediction. The recommendation system based on the DL prediction model using clinical data provided would assist physicians in decision-making where long lists of laboratory tests are needed but not replace them. One flexible feature for utilizing the current model is that physicians could use a different threshold to get the varying level of laboratory test recommendations based on patient characteristics. For example, S-GOT/AST, and S-GPT/ALT are not essential to perform on every patient with hypertension. Nevertheless, the physician can increase the cut-off value to get more recommendations and order these tests if they suspect any symptoms of kidney problem.

4.2. Current Research Gap and Possibility

Laboratory testing is a crucial part of healthcare for screening, diagnosing, treating, and monitoring disease progression. The use of laboratory tests is growing fast, therefore, selecting appropriate testing is important to improve patient safety by reducing diagnosis error [25,26]. Studies have pointed out the higher rate of the under- and over-utilization problem, indicating a perceivable gap for quality improvement [26,27,28]. There is inadequate evidence and a research gap to support appropriate laboratory test ordering. Selecting appropriate tests for an individual patient is challenging for physicians. The probable influence of hidden interaction between laboratory tests has been overlooked [29]. Physicians usually use their medical knowledge, judgment, and experience to ensure potential tests for patient diagnosis and management. Furthermore, physicians rely on computerized provider order entry (CPOE) rule-based decision support tools to order laboratory tests. Actual benefits of these systems for the appropriateness of laboratory test ordering remain unclear [30]. Utilizing deep learning models for laboratory data integration and analyses will offer an enormous opportunity to improve diagnostic value [31]. As the number of diagnostic tests is steadily increasing, and their ordering pattern has been changing over time, it is essential to recognize and extract valuable information from numerous laboratory data that may help clinical practice. A deep learning model can forge a practical path of ensuring appropriate laboratory test utilization based on previously unrecognised patterns.

4.3. Clinical Implications

Inappropriate laboratory test ordering is a neglected issue which hampers patient safety and unnecessarily increases costs [12]. Inappropriate laboratory test ordering appears to be a common problem across all high-resource healthcare settings [32,33]. This current study’s findings recommend a more wide-scale implementation of this system on national-level hospitals to reduce over- and under-utilization problems. Our recommendation model is rigorous and developed based on clinical practice data, therefore, it can be implemented at international level by adjusting thresholds. Various machine-learning models are available for laboratory test results, but no study focuses on laboratory test ordering. Over-utilization of laboratory tests has been reported in many studies [34,35]. However, in our current study, we focused extensively on pre-analytical laboratory test ordering error and developed a DL model that recognizes laboratory data patterns and provides personalized recommendations. Recently, Wright et al. [36] developed association rule mining for identifying a possible association between disease and laboratory tests results using 272,749 diseases, and 11,801,068 laboratory results. However, their automated technique for predicting test ordering has significant limitations, while our model predicted laboratory tests with a convincing performance.
Ideally, this DL-based clinical laboratory test ordering system may be integrated into the CPOE, and it will work in the background providing real-time, personalized laboratory test recommendations based on the patients’ available clinical histories. Consequently, physicians do not need to consider the complex and hidden interconnection between laboratory tests, guidelines, cost display, and fear of missing orders. Indeed, the system can improve clinical practice. Another benefit of using this model is collecting more patient data over time and improving its predictive ability. Our model only used four types of variables. These clinical variables are routinely recorded during visits to the hospital, therefore, these variables are always available in the EHR. Moreover, our model was designed to handle multiple outcomes at the same time.

4.4. Limitations

Our study has several limitations. First, we did not consider the temporal dimension in our current study. However, few laboratory tests depend on the temporal dimension. Second, our study did not include procedure code. However, some laboratory tests are connected to the procedures. Third, we did not externally validate our findings. Therefore, our findings might vary with other datasets. Although, we internally validated our model.

4.5. Future Prospects

The higher recall and precision at varying cut-off value levels suggest that our model can be integrated into the EHR to recommend personalized laboratory tests. We anticipate that the role of a DL model for suggesting laboratory tests will become a favoured and helpful tool in the near future. This approach will be a new pathway of ensuring improved patient care. In future, we have a plan to include more variables, such as procedure codes, to predict laboratory tests. This is because some laboratory tests fully depend on procedures. Moreover, we plan to add more data to improve our current study’s performance because the DL model works even better when the data volume is even higher. We will try to cover all laboratory tests as well as laboratory procedures in the future model. Furthermore, inpatients’ data and patients’ clinical notes will be considered to make our recommendation system even better.

5. Conclusions

We developed and validated the DNN model for recommending personalized laboratory tests using simple variables available in the EHR. The findings of our study show that this model has immense potential to predict appropriate laboratory tests, and help to reduce over- and under-utilization. Therefore, integrating this model into electronic health records systems may facilitate optimal test selection and reduce under- and over-utilization problems. However, the DL model is often referred to as a “black box” due to the lack of interpretability of findings [37,38]. Future studies are needed to evaluate the effectiveness of DL to predict laboratory tests.

Supplementary Materials

Author Contributions

Conceptualization, M.M.I. and Y.-C.L. Methodology, M.M.I.; software, M.M.I. Validation, M.M.I., and Y.-C.L. Formal analysis, T.N.P. Investigation, H.-C.Y.; resources, T.N.P.; data curation, T.N.P.; writing—original draft preparation, M.M.I.; writing—review and editing, Y.-C.L.; visualization, M.M.I.; supervision, Y.-C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is sponsored, in part, by the Ministry of Education (MOE) (Grant MOE DP2-110-21121-01-A-01) and the Ministry of Science and Technology (MOST) (Grant MOST 110-2321-B-038-002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Agarwal, R. Quality-improvement measures as effective ways of preventing laboratory errors. Lab. Med. 2014, 45, e80–e88. [Google Scholar] [CrossRef] [Green Version]
  2. Aesif, S.W.; Parenti, D.M.; Lesky, L.; Keiser, J.F. A cost-effective interdisciplinary approach to microbiologic send-out test use. Arch. Pathol. Lab. Med. 2015, 139, 194–198. [Google Scholar] [CrossRef]
  3. Vinker, S.; Kvint, I.; Erez, R.; Elhayany, A.; Kahan, E. Effect of the characteristics of family physicians on their utilisation of laboratory tests. Br. J. Gen. Pract. 2007, 57, 377–382. [Google Scholar] [PubMed]
  4. O’Sullivan, J.W.; Stevens, S.; Hobbs, F.R.; Salisbury, C.; Little, P.; Goldacre, B.; Bankhead, C.; Aronson, J.K.; Perera, R.; Heneghan, C. Temporal trends in use of tests in UK primary care, 2000-15: Retrospective analysis of 250 million tests. BMJ 2018, 363, k4666. [Google Scholar] [CrossRef] [Green Version]
  5. Hickner, J.; Thompson, P.J.; Wilkinson, T.; Epner, P.; Shaheen, M.; Pollock, A.M.; Lee, J.; Duke, C.C.; Jackson, B.R.; Taylor, J.R. Primary care physicians’ challenges in ordering clinical laboratory tests and interpreting results. J. Am. Board Fam. Med. 2014, 27, 268–274. [Google Scholar] [CrossRef] [Green Version]
  6. Zhi, M.; Ding, E.L.; Theisen-Toupal, J.; Whelan, J.; Arnaout, R. The landscape of inappropriate laboratory testing: A 15-year meta-analysis. PLoS ONE 2013, 8, e78962. [Google Scholar] [CrossRef] [Green Version]
  7. Ferraro, S.; Panteghini, M. The role of laboratory in ensuring appropriate test requests. Clin. Biochem. 2017, 50, 555–561. [Google Scholar] [CrossRef]
  8. Schumacher, L.D.; Jäger, L.; Meier, R.; Rachamin, Y.; Senn, O.; Rosemann, T.; Markun, S. Trends and between-Physician Variation in Laboratory Testing: A Retrospective Longitudinal Study in General Practice. J. Clin. Med. 2020, 9, 1787. [Google Scholar] [CrossRef]
  9. Huber, C.A.; Nagler, M.; Rosemann, T.; Blozik, E.; Näpflin, M.; Markun, S. Trends in Micronutrient Laboratory Testing in Switzerland: A 7-Year Retrospective Analysis of Healthcare Claims Data. Int. J. Gen. Med. 2020, 13, 1341. [Google Scholar] [CrossRef] [PubMed]
  10. Lippi, G.; Bovo, C.; Ciaccio, M. Inappropriateness in laboratory medicine: An elephant in the room? Ann. Transl. Med. 2017, 5, 82. [Google Scholar] [CrossRef] [Green Version]
  11. Ganguli, I.; Simpkin, A.L.; Lupo, C.; Weissman, A.; Mainor, A.J.; Orav, E.J.; Rosenthal, M.B.; Colla, C.H.; Sequist, T.D. Cascades of care after incidental findings in a US national survey of physicians. JAMA Netw. Open 2019, 2, e1913325. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Vrijsen, B.; Naaktgeboren, C.; Vos, L.; van Solinge, W.; Kaasjager, H.; Ten Berg, M. Inappropriate laboratory testing in internal medicine inpatients: Prevalence, causes and interventions. Ann. Med. Surg. 2020, 51, 48–53. [Google Scholar] [CrossRef] [PubMed]
  13. Basuita, M.; Kamen, C.L.; Ethier, C.; Soong, C. Reducing Inappropriate Laboratory Testing in the Hospital Setting: How Low Can We Go? JCOM 2020, 27. [Google Scholar] [CrossRef]
  14. Mrazek, C.; Simundic, A.-M.; Salinas, M.; von Meyer, A.; Cornes, M.; Bauçà, J.M.; Nybo, M.; Lippi, G.; Haschke-Becher, E.; Keppel, M.H. Inappropriate use of laboratory tests: How availability triggers demand–Examples across Europe. Clin. Chim. Acta 2020, 505, 100–107. [Google Scholar] [CrossRef] [PubMed]
  15. O’Sullivan, J.W.; Albasri, A.; Nicholson, B.D.; Perera, R.; Aronson, J.K.; Roberts, N.; Heneghan, C. Overtesting and undertesting in primary care: A systematic review and meta-analysis. BMJ Open 2018, 8, e018557. [Google Scholar] [CrossRef] [Green Version]
  16. Maillet, É.; Paré, G.; Currie, L.M.; Raymond, L.; de Guinea, A.O.; Trudel, M.-C.; Marsan, J. Laboratory testing in primary care: A systematic review of health IT impacts. Int. J. Med Inform. 2018, 116, 52–69. [Google Scholar] [CrossRef] [Green Version]
  17. Delvaux, N.; Van Thienen, K.; Heselmans, A.; de Velde, S.V.; Ramaekers, D.; Aertgeerts, B. The effects of computerized clinical decision support systems on laboratory test ordering: A systematic review. Arch. Pathol. Lab. Med. 2017, 141, 585–595. [Google Scholar] [CrossRef] [Green Version]
  18. Khalifa, M.; Khalid, P. Reducing unnecessary laboratory testing using health informatics applications: A case study on a tertiary care hospital. Procedia Comput. Sci. 2014, 37, 253–260. [Google Scholar] [CrossRef] [Green Version]
  19. Fang, D.Z.; Sran, G.; Gessner, D.; Loftus, P.D.; Folkins, A.; Christopher, J.Y.; Shieh, L. Cost and turn-around time display decreases inpatient ordering of reference laboratory tests: A time series. BMJ Qual. Saf. 2014, 23, 994–1000. [Google Scholar] [CrossRef]
  20. Eaton, K.P.; Levy, K.; Soong, C.; Pahwa, A.K.; Petrilli, C.; Ziemba, J.B.; Cho, H.J.; Alban, R.; Blanck, J.F.; Parsons, A.S. Evidence-based guidelines to eliminate repetitive laboratory testing. JAMA Intern. Med. 2017, 177, 1833–1839. [Google Scholar] [CrossRef]
  21. Schubart, J.R.; Fowler, C.E.; Donowitz, G.R.; Connors, A.F., Jr. Algorithm-based decision rules to safely reduce laboratory test ordering. Stud. Health Technol. Inform. 2001, 84, 523–527. [Google Scholar]
  22. Atique, S.; Hsieh, C.-H.; Hsiao, R.-T.; Iqbal, U.; Nguyen, P.A.A.; Islam, M.M.; Li, Y.-C.J.; Hsu, C.-Y.; Chuang, T.-W.; Syed-Abdul, S. Viral warts (Human Papilloma Virus) as a potential risk for breast cancer among younger females. Comput. Methods Programs Biomed. 2017, 144, 203–207. [Google Scholar] [CrossRef]
  23. Tsai, C.-Y.; Yang, H.-C.; Islam, M.; Hsieh, W.-S.; Juan, S.-H.; Chen, J.-C.; Khan, H.A.A.; Jian, W.-S. Psychotropic medications prescribing trends in adolescents: A nationwide population-based study in Taiwan. Int. J. Qual. Health Care 2017, 29, 861–866. [Google Scholar] [CrossRef] [Green Version]
  24. Hsing, A.W.; Ioannidis, J.P. Nationwide population science: Lessons from the Taiwan national health insurance research database. JAMA Intern. Med. 2015, 175, 1527–1529. [Google Scholar] [CrossRef]
  25. Luo, Y.; Szolovits, P.; Dighe, A.S.; Baron, J.M. Using machine learning to predict laboratory test results. Am. J. Clin. Pathol. 2016, 145, 778–788. [Google Scholar] [CrossRef] [Green Version]
  26. National Academies of Sciences, Engineering, and Medicine. Improving Diagnosis in Health Care; National Academies Press: Washington, DC, USA, 2015.
  27. Janssens, P.M. Managing the demand for laboratory testing: Options and opportunities. Clin. Chim. Acta 2010, 411, 1596–1602. [Google Scholar] [CrossRef]
  28. Fryer, A.A.; Smellie, W.S.A. Managing demand for laboratory tests: A laboratory toolkit. J. Clin. Pathol. 2013, 66, 62–72. [Google Scholar] [CrossRef] [Green Version]
  29. Yeh, D.D. A clinician’s perspective on laboratory utilization management. Clin. Chim. Acta 2014, 427, 145–150. [Google Scholar] [CrossRef]
  30. Delvaux, N.; Piessens, V.; De Burghgraeve, T.; Mamouris, P.; Vaes, B.; Vander Stichele, R.; Cloetens, H.; Thomas, J.; Ramaekers, D.; De Sutter, A. Clinical decision support improves the appropriateness of laboratory test ordering in primary care without increasing diagnostic error: The ELMO cluster randomized trial. Implement. Sci. 2020, 15, 1–10. [Google Scholar] [CrossRef] [PubMed]
  31. Louis, D.N.; Gerber, G.K.; Baron, J.M.; Bry, L.; Dighe, A.S.; Getz, G.; Higgins, J.M.; Kuo, F.C.; Lane, W.J.; Michaelson, J.S. Computational pathology: An emerging definition. Arch. Pathol. Lab. Med. 2014, 138, 1133–1138. [Google Scholar] [CrossRef]
  32. Hauser, R.G.; Shirts, B.H. Do we now know what inappropriate laboratory utilization is? An expanded systematic review of laboratory clinical audits. Am. J. Clin. Pathol. 2014, 141, 774–783. [Google Scholar] [CrossRef]
  33. Akan, P.; Cimrin, D.; Ormen, M.; Kume, T.; Ozkaya, A.; Ergor, G.; Abacioglu, H. The inappropriate use of HbA1c testing to monitor glycemia: Is there evidence in laboratory data? J. Eval. Clin. Pract. 2007, 13, 21–24. [Google Scholar] [CrossRef]
  34. Hawkins, R.C. Potentially inappropriate repeat laboratory testing in inpatients. Clin. Chem. 2006, 52, 784–785. [Google Scholar] [CrossRef]
  35. Weydert, J.A.; Nobbs, N.D.; Feld, R.; Kemp, J.D. A simple, focused, computerized query to detect overutilization of laboratory tests. Arch. Pathol. Lab. Med. 2005, 129, 1141–1143. [Google Scholar] [CrossRef] [PubMed]
  36. Wright, A.; Chen, E.S.; Maloney, F.L. An automated technique for identifying associations between medications, laboratory results and problems. J. Biomed. Inform. 2010, 43, 891–901. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Castelvecchi, D. Can we open the black box of AI? Nat. News 2016, 538, 20. [Google Scholar] [CrossRef] [Green Version]
  38. Buhrmester, V.; Münch, D.; Arens, M. Analysis of explainers of black box deep neural networks for computer vision: A survey. arXiv 2019, arXiv:191112116. [Google Scholar]
Figure 1. Selection of disease using the first three-digit code.
Figure 1. Selection of disease using the first three-digit code.
Diagnostics 11 00990 g001
Figure 2. Selection of five digits ATC code.
Figure 2. Selection of five digits ATC code.
Diagnostics 11 00990 g002
Figure 3. Illustrations of multi-label classification. X is the data set in which feature vectors represent patients C1–Cn; n is the number of patients; F1–Fm are variables; m is the number of variables. Y is the label vector; L1–L4 represents the number of lab tests.
Figure 3. Illustrations of multi-label classification. X is the data set in which feature vectors represent patients C1–Cn; n is the number of patients; F1–Fm are variables; m is the number of variables. Y is the label vector; L1–L4 represents the number of lab tests.
Diagnostics 11 00990 g003
Figure 4. An architecture of the proposed deep learning model.
Figure 4. An architecture of the proposed deep learning model.
Diagnostics 11 00990 g004
Figure 5. Training and validation loss of the deep learning model.
Figure 5. Training and validation loss of the deep learning model.
Diagnostics 11 00990 g005
Figure 6. Receiver operating characteristic (ROC) curves of the deep learning model for predicting laboratory tests. The range of AUROC was between 0.76 and 1 (Supplementary Table S1). The highest AUROC 1 was HIV viral loading test and the lowest AUROC 0.76 was for CBC-Ⅲ (WBC, RBC, HB, HCT, and MCV).
Figure 6. Receiver operating characteristic (ROC) curves of the deep learning model for predicting laboratory tests. The range of AUROC was between 0.76 and 1 (Supplementary Table S1). The highest AUROC 1 was HIV viral loading test and the lowest AUROC 0.76 was for CBC-Ⅲ (WBC, RBC, HB, HCT, and MCV).
Diagnostics 11 00990 g006
Figure 7. Total number of laboratory tests with their AUROC range.
Figure 7. Total number of laboratory tests with their AUROC range.
Diagnostics 11 00990 g007
Figure 8. Evaluation of the performance of the deep learning model for predicting laboratory tests, (A) system always give exact recommendation for laboratory tests with a cut off value 0.35, (B) system might give extra recommendation of laboratory tests with a cut off value 0.30; however all laboratory recommendations are correct. Physicians can adjust cut-off value if additional tests are required. Decreasing cut-off value will increase the number of laboratory tests recommendation.
Figure 8. Evaluation of the performance of the deep learning model for predicting laboratory tests, (A) system always give exact recommendation for laboratory tests with a cut off value 0.35, (B) system might give extra recommendation of laboratory tests with a cut off value 0.30; however all laboratory recommendations are correct. Physicians can adjust cut-off value if additional tests are required. Decreasing cut-off value will increase the number of laboratory tests recommendation.
Diagnostics 11 00990 g008
Table 1. The performance of DL model based on varying cut-offs for clinical laboratory test prediction.
Table 1. The performance of DL model based on varying cut-offs for clinical laboratory test prediction.
Cut-OffPrecisionRecallF1 ScoreHamming Loss
≥0.010.240.960.370.082
≥0.050.320.890.460.038
≥0.100.390.830.520.025
≥0.150.440.760.550.019
≥0.200.480.720.560.016
≥0.250.520.660.570.014
≥0.300.550.610.570.013
≥0.350.570.570.550.012
≥0.400.610.510.540.011
≥0.450.630.470.510.011
≥0.500.650.420.480.011
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Islam, M.M.; Poly, T.N.; Yang, H.-C.; Li, Y.-C. Deep into Laboratory: An Artificial Intelligence Approach to Recommend Laboratory Tests. Diagnostics 2021, 11, 990. https://doi.org/10.3390/diagnostics11060990

AMA Style

Islam MM, Poly TN, Yang H-C, Li Y-C. Deep into Laboratory: An Artificial Intelligence Approach to Recommend Laboratory Tests. Diagnostics. 2021; 11(6):990. https://doi.org/10.3390/diagnostics11060990

Chicago/Turabian Style

Islam, Md. Mohaimenul, Tahmina Nasrin Poly, Hsuan-Chia Yang, and Yu-Chuan (Jack) Li. 2021. "Deep into Laboratory: An Artificial Intelligence Approach to Recommend Laboratory Tests" Diagnostics 11, no. 6: 990. https://doi.org/10.3390/diagnostics11060990

APA Style

Islam, M. M., Poly, T. N., Yang, H. -C., & Li, Y. -C. (2021). Deep into Laboratory: An Artificial Intelligence Approach to Recommend Laboratory Tests. Diagnostics, 11(6), 990. https://doi.org/10.3390/diagnostics11060990

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop