Next Article in Journal
The Application of Radiomics and AI to Molecular Imaging for Prostate Cancer
Next Article in Special Issue
Artificial Intelligence-Supported Ultrasonography in Anesthesiology: Evaluation of a Patient in the Operating Theatre
Previous Article in Journal
Special Issue “Personalized Medicine in Blood Disease of Children”
Previous Article in Special Issue
Individualized Perioperative Hemodynamic Management Using Hypotension Prediction Index Software and the Dynamics of Troponin and NTproBNP Concentration Changes in Patients Undergoing Oncological Abdominal Surgery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Ultrasonographic Applications of Novel Technologies and Artificial Intelligence in Critically Ill Patients

1
Medica Co., Ltd. (Upper Silesian School of Ultrasonography), 41-500 Chorzów, Poland
2
Collegium Medicum, Jan Kochanowski University (JKU), 25-317 Kielce, Poland
3
Municipal Hospital Co., Ltd., 41-703 Ruda Śląska, Poland
4
Collegium Medicum, WSB University, 41-300 Dąbrowa Górnicza, Poland
5
Department of Anaesthesiology and Critical Care, School of Medicine with the Division of Dentistry, Medical University of Silesia, 41-808 Zabrze, Poland
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2024, 14(3), 286; https://doi.org/10.3390/jpm14030286
Submission received: 17 January 2024 / Revised: 21 February 2024 / Accepted: 26 February 2024 / Published: 7 March 2024
(This article belongs to the Special Issue Personalized Medicine in Anesthesia and Anesthetics)

Abstract

:
The diagnostic process in Intensive Care Units has been revolutionized by ultrasonography and accelerated by artificial intelligence. Patients in critical condition are often sonoanatomically challenging, with time constraints being an additional stress factor. In this paper, we describe the technology behind the development of AI systems to support diagnostic ultrasound in intensive care units. Among the AI-based solutions, the focus was placed on systems supporting cardiac ultrasound, such as Smart-VTI, Auto-VTI, SmartEcho Vue, AutoEF, Us2.ai, and Real Time EF. Solutions to assist hemodynamic assessment based on the evaluation of the inferior vena cava, such as Smart-IVC or Auto-IVC, as well as to facilitate ultrasound assessment of the lungs, such as Smart B-line or Auto B-line, and to help in the estimation of gastric contents, such as Auto Gastric Antrum, were also discussed. All these solutions provide doctors with support by making it easier to obtain appropriate diagnostically correct ultrasound images by automatically performing time-consuming measurements and enabling real-time analysis of the obtained data. Artificial intelligence will most likely be used in the future to create advanced systems facilitating the diagnostic and therapeutic process in intensive care units.

1. Introduction

Ultrasound has a particular application in intensive care. Since the sonoanatomical conditions in patients treated in intensive care units (ICU) are far from ideal, and ultrasound (US) imaging is often performed under time pressure, physicians are increasingly more likely to use artificial intelligence (AI). In particular, the use of AI in intensive care has been discussed very often in both the literature and the media in the last 2–3 years. Modern technologies have led to progress in the development of medical equipment dedicated to anesthesiology and intensive care [1]. AI has been widely used in both these fields, making it significantly easier to interpret the obtained images and providing doctors with rapid feedback on the patient’s clinical situation [2,3].
Point-of-care ultrasound (POCUS) has a wide diagnostic application in intensive care, making it easier to use and interpret such imaging modalities as transthoracic and transesophageal echocardiography, lung ultrasound, ultrasound of a patient in shock, and patients with other life-threatening conditions. Due to the complexity of ultrasound assessment in intensive care, the physician’s purpose of using AI in this setting is to help assess the patient’s condition based on US imaging of various systems and organs and assess the patient’s hemodynamic profile [4,5,6]. The literature has shown that AI-guided POCUS will most likely increase the efficiency and effectiveness of ultrasound imaging [4,7]. AI-supported point-of-care protocols also have particular educational applications, enabling young medical students to improve ultrasound techniques and increasing the identifiability of anatomical structures, which is necessary to perform procedures [8].
Excellent knowledge of normal, topographic, and pathological anatomy is essential for a correct assessment of ultrasound images. High inter-patient anatomical variability and often difficult sonoanatomical conditions may discourage less experienced anesthesiologists from performing US [3]. Point-of-care protocols require excellent technique in operating an ultrasonic transducer from the operator due to time constraints and the schematic nature of this type of examination.
In this complex setting, AI comes to the aid of clinicians [9]. Depending on the type of software, it helps detect key structures for performing a specific medical procedure, such as cannulation, finding the optimal ultrasound plane for imaging of the heart, lungs, or abdominal organs [4,10] as well as setting the appropriate image quality, and performing measurements, including complex Doppler measurements [11]. At the same time, it significantly reduces the time needed to learn how to perform imaging and the time to acquire clinically significant images, and it improves the accuracy of the obtained results [2,10,12,13].

2. Selected Innovative AI-Based US Technologies and Solutions Intended for Critically Ill Patients

Intensive care ultrasonography (POCUS) is a tool that can provide abundant clinically relevant information about a patient’s condition. The doctors attending the patient change, which, combined with the diversity of medical staff in terms of ultrasound imaging skills, may lead to variable and often incomplete observations and conclusions. A typical POC examination is often performed under time pressure. AI-based solutions prove extremely helpful in this regard and are very dynamically entering everyday clinical practice. They allow for obtaining more objective real-time results than those obtained by different operators [14]. Similar to regional anesthesiology, these solutions are based on machine learning (ML), deep learning (DL) and deep convolutional neutral network (DCNN), U-Net, and big data. These are information technologies used in the processing and very detailed interpretation of images, including medical images. In the case of DL, the created software is intended to imitate and simulate human thinking in terms of the processed information and is an important component of ML, which together constitute artificial intelligence. ML, in turn, is the machine’s ability to automatically learn and improve the interpretation of images [15]. The key technologies discussed in this paper include the following.
SmartVTI (Mindray Medical International Limited, Shenzhen, China) hereinafter Mindray), AutoVTI (General Electric, Boston, MA, USA), US2.AI. Cardiac Output Workflow (EchoNous Inc., Redmond Washington, DC, USA) or LVivo Seamless (Philips Medical Systems International, Best, Holandia) are examples of AI-based tools offered by two different manufacturers, which are used in ultrasound of patients in a life-threatening condition. These semi-automatic solutions make it possible to calculate two key parameters for hemodynamic monitoring and provide important prognostic data for a critically ill patient. These two parameters are the automatic measurement of stroke volume (SV) in mL and the cardiac output (CO), which is the product of the heart rate (HR) and the SV [15,16,17]. Typical and manual examination of these two measurements requires at least two ultrasound views (PLAX—parasternal long axis view), as well as LVOT (left ventricular outflow tract) measurement, and Doppler spectrum contouring by setting the sampling gate parallel to the LVOT blood flow from the five-chamber view (5CH). The inter-operator results may differ due to incorrect placement of the Doppler gate and incorrect selection of the insonation angle, which may consequently lead to overestimation of the velocity time integral (VTI) [18]. Considering the fact that the SV is the product of LVOT and VTI, the final value of the stroke volume may vary significantly depending on the operator’s skill and precision. The sequence of steps in the manual approach is shown in Figure 1.
When using AI-based tools, the operator should perform only one manual measurement, i.e., LVOT in the PLAX view. The second and last step is to obtain a 5CH view and, through “one-click” (SmartVTI for Mindray or AutoVTI for GE devices), the device will automatically set the Doppler gate in the left ventricular outflow tract (LVOT) and outline the Doppler spectrum by tracing the velocity in order to calculate the VTI in real time. With data in the form of the LVOT area and VTI as well as the heart rate (HR), the ultrasound device is able to automatically calculate the SV and CO parameters in real time, presenting the data in the form of trending graphs (Figure 2).
The calculations consist of simple mathematical operations using the following formulas [19]:
SV (mL/cycle) = LVOT Area (cm2) × VTI/cycle (cm/cycle)
CO (L/min) = SV (L/cycle) × HR (cycle/min)
Additionally, some manufacturers of ultrasound equipment have access to technology that allows the assessment of views obtained by the operator during an examination. They bear in mind the fact that the above measurements are reliable only when appropriate PLAX and 5CH images are obtained, and the device provides the operator with a graphic “prompt” on whether the obtained image is acceptable or whether it should be corrected. By properly selecting colors, the operator knows whether a given view is appropriate or whether it requires more involvement. Only when the obtained images are accepted by the AI algorithm are the automatic measurements taken [20]. An example of SmartEchoVue (Mindray), Us2.ai and AI TRO (EchoNous Inc.), Butterfly ScanLab (Butterfly Network, Inc.) or LVivo Seamless (Philips) technology enabling the correction of appropriate views is shown in Figure 3.
In the case of solutions proposed by Philips in the LVivo Seamless technology, the device itself selects the most optimal projections obtained during an ultrasound scan and uses them to present cardiac measurements, including SV [21] (Figure 4).
AI TRIO (EchoNous, Inc.) and Butterfly ScanLab, which offer real-time support during an ultrasound scan, are yet another solution implemented by EchoNous, Inc., and Butterfly Network, Inc. (Boston, MA, USA), which allows even beginners to quickly make progress in performing echocardiography. They help the operator to identify quickly and very accurately individual cardiac structures and assess the quality of and initially qualify the obtained images. This solution is presented in Figure 5 and Figure 6.
AutoEF (Mindray)/Real Time EF (GE)/ Us2.ai (EchoNous Inc.)/LVivo EF (Philips) is another tool to assist clinicians in their work, allowing the calculation of the ejection fraction (EF) in real time based on AI. The assessment of the left ventricular (LV) EF is an integral part of echocardiographic evaluation. Despite not being error-free, it is one of the basic criteria for therapeutic decisions and a prognostic marker. EF assessment allows for a quantitative assessment of the blood volume ejected from the left ventricle during each heartbeat. In the traditional and manual approach, it is a multi-stage process that includes both systolic and diastolic measurements of the left ventricle. Measurements can be made in two views: 4CH (four chamber view) and 2CH (two chamber view) as shown in Figure 7.
Based on the obtained measurements, the ultrasound machine uses the built-in calculation package to estimate the EF according to the following formula:
EF = (End Diastolic Volume − End Systolic Volume)/End Diastolic Volume × 100
This measurement is based on the Simpson method, i.e., the biplane method of disk summation as shown in Figure 8.
Manual outlining of the LV endocardium may be very time consuming and burdened by inter-operator measurement errors, not to mention the significant time needed to perform the scan itself. In the intensive care and according to the POCUS philosophy, often, only a quantitative method is used to assess global LV dysfunction and thereby the patient’s general status. AutoEF or RealTime EF may be a tool to solve many of the abovementioned problems related to EF assessment (Figure 9).
The operator’s main task is to obtain an apical four chamber view (A4CH). Using AI algorithms, ML and DL in particular, the ultrasound device first classifies the image obtained by the operator by tracing individual ventricular wall segments both in the end-diastolic and end-systolic phases, while assessing the maximum and minimum LV volume. If the view and the correct image have been accepted by the ultrasound machine, the most optimal views are selected in one cycle, and the LVEF is automatically calculated. The calculated parameters are displayed on the ultrasound screen. Each time, the operator has the opportunity to correct the outlined LV endocardium if necessary.
The technology of the automatic measurement of the inferior vena cava (IVC), referred to by various manufacturers as SmartIVC (Mindray) and AutoIVC (GE), is an additional tool to support the patient’s assessment and deliver clinically relevant information (Figure 10). The IVC width and collapsibility are indirect indicators of myocardial preload and CVP estimation. Similar to LVEF measurement, ultrasonographic IVC assessment can also be highly subjective. The final dimension of the inferior vena cava is influenced by, among other things, the operator’s skill and, above all, the unstable position of this vessel during respiratory movements. During spontaneous breathing, the IVC moves on average 21.7 mm vertically and 3.9 mm horizontally. Furthermore, the actual vertical axis of the IVC in the human body is not 90° but an average of 115°. All this means that the test is not easy to interpret, and it should take into account these limitations [24,25].
According to the 2015 guidelines of the American Society of Echocardiography (ASE) [26], the measurement should be performed with the patient in a supine position, as the IVC diameter and shape may change depending on the patient’s position, at a distance of 1–2 cm from the opening of the hepatic vein, perpendicular to the long axis of the vessel in B-Mode. In clinical practice, M-Mode is also used to observe IVC respiratory variability [24]. The traditional and manual diagnostic process is presented in Figure 11.
The semi-auto SmartIVC and AutoIVC provide quick and simple real-time IVC measurements without the need to manually perform the individual stages of the examination described in Figure 11.
After obtaining a correct view in the long axis of the IVC and activating the SmartIVC or AutoIVC button on the US control panel, the algorithm automatically determines the anatomical M-line, which is placed exactly at the right site, where the measurements are made manually. Both the angle and the position of the M-line are adjusted in real time using a special tracing algorithm. If the image changes due to the mobility of the IVC described earlier, the M-line automatically adjusts so that it is always perpendicular to IVC axis. The user can always adjust the position and angle of the anatomical M-line. Owing to the automatic detection of the IVC walls, the presented results show its maximum and minimum dimensions and the Caval Index (CI). It is calculated according to the following formula:
Caval Index (CI) = (Maximum diameter (expiratory) − Minimum diameter (inspiratory))/Maximum diameter (expiratory).
In the case of a mechanically ventilated patient, the DI (Distensibility Index) is presented instead of the CI, calculated according to the following formula:
Distensibility Index (DI) = (Maximum diameter (expiratory) − Minimum diameter (inspiratory))/Minimum diameter (inspiratory).
Furthermore, in addition to the CI and DI, the technology allows for presenting the trend lines in accordance with the change in collapsibility, as shown in Figure 12.
The Inferior Vena Cava Distensibility Index (IVCDI) is a parameter commonly used in critical care units to assess the fluid responsiveness and intravascular volume status in ventilated patients. It involves measuring changes in the diameter of the inferior vena cava (IVC) during the respiratory cycle. The IVC in deeply sedated or paralyzed patients being passively ventilated with positive pressure has a very different pattern in respiratory variation compared to that of spontaneously breathing patients. In the absence of central hypovolemia, the increase in the internal IVC pressure that has been transmitted from the right atrium exceeds the increase in the external intra-abdominal pressure. This pressure gradient will distend the IVC; so, its size will increase on inspiration. This does not occur in a patient who is fluid unresponsive. When the caval venous system is congested and closer to its maximal size, there will be minimal, if any, distention of the IVC with inspiration. These phenomena can be quantified by the “IVC Distensibility Index”.
The Vena Cava Distensibility Index and distensibility variability can be calculated with the following formulas: IVC-DI = (max diameter–min diameter)/(min diameter) × 100, IVC-DV = [(maximum diameter − minimum diameter)/(mean diameter)] × 100). The cut off of the distensibility index for fluid responsiveness is between 15 and 20%, with maximum accuracy when Vt ≥ 8 mL/Kg or PEEP ≤ 5 cm H2O.
However, to reiterate, the change in IVC size with positive pressure mechanical ventilation is only observed consistently in patients who have no inspiratory effort and are being ventilated passively. These are patients who are either deeply sedated or paralyzed [27,28,29,30,31].
Airway ultrasound is becoming a common, accessible, quick, and easy method for assessing the upper airways. Its main goal, especially in the case of intensive care patients, is to identify anatomical structures such as thyroid cartilage, epiglottis, cricoid cartilage, cricothyroid membrane, tracheal cartilages, and esophagus, determine the required size of endotracheal tubes (ETT), aid potentially difficult intubation, and help perform a percutaneous tracheostomy [32]. Research on the use of AI in airway ultrasound is currently underway, pointing to the enormous potential of this technology [32]. Due to its specificity, airway US is also related to lung ultrasound, in which AI is already used. In addition to the previously mentioned assessment of the myocardium and the inferior vena cava, AI technologies are also utilized to assess the respiratory system. Since 1992, when Professor Daniel Lichtenstein published an article on lung ultrasound, this discipline has become increasingly appreciated by clinicians and is increasingly used in critically ill patients. Lung ultrasound is used to interpret artifacts, the presence or absence of which may be of key importance in the clinical assessment. B-lines are one of the key findings assessed in lung ultrasound. The presence of these vertical artifacts excludes pneumothorax on the one hand but is a marker of interstitial fluid accumulation on the other hand. The new guidelines (February 2023) [33] for the use of lung ultrasound not only once again clearly confirm that AI has a wide application in many fields of medicine but also point to its powerful potential in lung diagnosis. Combining technology and diagnostic imaging may prove extremely useful as the first screening in an intensive care unit [33]. Conventional lung ultrasound focuses on the assessment of horizontal and vertical artifacts and involves their visual assessment. A convex transducer is used for fluid components, with the operator focused on the quantitative and qualitative B-line assessment on the ultrasound screen. As in the case of the previously described ultrasound scans, the assessment and interpretation of the obtained findings may vary depending on the operator’s skill and experience. Therefore, another AI-based technology may be of great support for the assessment of vertical US artifacts. This technology is referred to as Smart B-line (Mindray) and AutoB-line (GE), depending on the manufacturer. A doctor performing a lung scan using this technology positions the ultrasound transducer in exactly the same zones as during conventional imaging. The US machine can assign a diagnostic image to each scanned zone. There are 6, 8, and 12 scanning zones, the number of which is selected by the operator as described Figure 13.
Automatic identification of the pleural line, the rib acoustic shadow, and the presentation of the scanned zone for the presence of vertical artifacts (B-lines) is the primary support offered by this technology. Once the operator accepts the proposed zone, the B-lines in the selected zone are identified and automatically counted, and the distance between them is measured. In addition to these data, the ultrasound screen also displays information about the B-line area ratio for the entire zone (%), as shown in Figure 14.
Once the scan is complete, the operator can generate a report that will summarize the entire examination along with the scoring of each scanned zone, as shown in Figure 15.
More than vertical artifacts can be identified by AI-based technologies in LUS. The technology used in the latest solution from Butterfly Network, Inc. helps identify horizontal artifacts, i.e., the A-lines. As shown in Figure 16, Butterfly ScanLab software employed in the Butterfly iQ+ device identifies horizontal artifacts and the pleura itself during an ultrasound scan, presenting them in appropriate colors.
To summarize, AI-based technologies are used in the ultrasound assessment of the lungs, including patients with COVID-19, and can be very helpful, especially for a less experienced physician, eliminating inter-operator subjectivity of the assessment and providing a large amount of valuable clinical information. Additionally, they are very easy to use [34,35].
Auto Gastric Antrum (Mindray), i.e., automatic measurement of gastric contents, is the last technology supported by DL algorithms. In clinical practice, this test is part of the gastric POCUS concept, which is a non-invasive quantitative and qualitative assessment of the gastric contents. Qualitative measurement helps determine whether the stomach is empty or filled with solid or liquid content. Quantitative assessment involves estimating the gastric volume, which is calculated based on the antral cross-sectional area (antral CSA) [36]. This information is very useful for the assessment of the risk of aspiration, which in turn facilitates therapeutic decisions and influences the type of anesthesia or induction techniques. A convex transducer is used, and the patient is placed in two positions—on the back and on the right side. The sagittal view is used as a standard. The operator performs quantitative and qualitative assessment by manually outlining the gastric cardia in the patient placed on the right side. The obtained CSA dimension is directly proportional to the volume, which is ultimately obtained from the data in a mathematical table that takes into account the patient’s age. If the test is repeated several times, it is necessary to archive the data to assess the trend of changes. When Auto Gastric Antrum is used, the operator’s only task is to obtain the correct image accepted by the device. From this moment, the AI-guided ultrasound system will evaluate the image, locate the antrum and automatically outline the CSA, presenting the GA area in cm2, as shown in Figure 17.
Since the amount of gastric contents varies over time, it is possible to present the CSA area in the form of a trend line, which is also presented in Figure 17. Therefore, as in each technology described earlier, the automation of diagnostic measurements performed by the AI technology objectifies the obtained results, making them repeatable and avoiding inter-operator variability. In clinical practice, it has become an easy and very useful tool for assessing gastric contents.
Clarius PAL HD3 (Clarius Mobile Health, Vancouver, BC, Canada) is the last intensive-care ultrasonographic solution based on AI, not only in the context of software but also in the context of the optimal use of a single ultrasound transducer to examine the patient’s entire body (Figure 18). This solution allows the use of a single wireless transducer for all ultrasound examinations in an intensive care unit. One transducer, in the shape of a typical phased array transducer, presents an appropriate image as a linear, convex, or phased array transducer, depending on the site where the transducer is applied by the operator. In practice, it works in such a way that if the operator applies the transducer, for example, to the patient’s abdomen, an image typical of a convex transducer will automatically appear on the ultrasound screen, without the need to press any buttons. If the transducer is positioned around the cervical vessels, the image shape will change to the one acquired with a typical linear transducer. If the scan is performed to assess the heart, the presented image will automatically change into one typical of phased array transducers.
If we add to this functionality the fact that the manufacturer also offers the option of voice control for the key functions of the ultrasound system, this solution becomes very practical for the operator.
All the above mentioned AI-based technologies (Table 1) are still at an early stage of implementation, and confirming their usefulness requires extensive research to enable their practical clinical evaluation. Simplified interpretation of images (63%);
  • Facilitation of the learning process for ultrasound (59%);
  • Streamlined capture of images (47%);
  • Enhanced accuracy scanning (42%);
  • Quicker diagnosis for accelerated treatment (40%);
  • Higher efficiency in workflow (38%).
The benefits of using AI include not only support for beginner ultrasonographers but also for those highly experienced. Survey participants pointed to significant improvements in productivity, with features such as auto-labeling (76%) and daily diagnostic support (75%) of key importance. Most of all, AI reduces the scanning time, which allows a larger number of patients to be examined. Additionally, diagnostic AI support may be used to confirm the initial diagnosis established by experienced clinicians, thus accelerating therapeutic processes in an intensive care unit (Figure 19) [38].

3. Conclusions

Introduction of ultrasound to assess a patient’s health status was the first revolution in intensive care. Currently, as shown in this paper, artificial intelligence facilitates ultrasound imaging and measurements crucial for selected areas of intensive care, significantly reducing the workload of medical staff. Introduction of complex systems based on artificial intelligence, which will increase the diagnostic speed and effectiveness, as well as facilitate making complex clinical decisions, will most likely represent another milestone. The complexity of an intensive care unit, the instability and severity of patients’ condition, and the need to combine the medical knowledge of specialists in various medical fields will most likely make it impossible to create independent autonomous systems for assessing patient health status, but the emerging solutions will significantly accelerate the diagnostic and therapeutic path, improving the treatment outcomes of intensive care patients.

Author Contributions

Conceptualization, S.M. and W.G.; methodology and software S.M., W.G., M.G.-M. and M.W.; validation, H.M.; formal analysis, S.M., W.G., M.G.-M., M.W. and H.M.; investigation S.M. and W.G.; resources, S.M.; data curation, W.G., M.G.-M., M.W. and H.M.; writing—original draft preparation, S.M., W.G., M.G.-M. and M.W.; writing—review and editing, S.M., W.G., M.G.-M., M.W. and H.M.; visualization, S.M. and M.W.; supervision, H.M., M.W., W.G. and M.G.-M.; project administration, S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

Author Monika Gil-Mika was employed by the company Municipal Hospital Co., Ltd and Author Sławomir Mika was employed by the company Medica Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Pham, F.M. Artificial Intelligence-Supported Systems in Anesthesiology and Its Standpoint to Date—A Review. Open J. Anesthesiol. 2023, 13, 140–168. [Google Scholar] [CrossRef]
  2. Kuroda, Y.; Kaneko, T.; Yoshikawa, H.; Uchiyama, S.; Nagata, Y.; Matsushita, Y.; Hiki, M.; Minamino, T.; Takahashi, K.; Daida, H.; et al. Artificial intelligence-based point-of-care lung ultrasound for screening COVID-19 pneumoniae: Comparison with CT scans. PLoS ONE 2023, 18, e0281127. [Google Scholar] [CrossRef] [PubMed]
  3. Suissa, N.; Jeffries, S.D.; Song, K.; Botros, S.; Hemmerling, T.M. Utilizing Deep Learning to Identify an Ultrasound-Guided Nerve Block Target Zone. Available online: https://arinex.com.au/EMBC/pdf/full-paper_26.pdf (accessed on 1 January 2024).
  4. Maheshwarappa, H.M.; Mishra, S.; Kulkarni, A.V.; Gunaseelan, V.; Kanchi, M. Use of Handheld Ultrasound Device with Artificial Intelligence for Evaluation of Cardiorespiratory System in COVID-19. Indian J. Crit. Care Med. 2021, 25, 524–527. [Google Scholar] [CrossRef] [PubMed]
  5. Ying, F.; Chen, S.; Pan, G.; He, Z. Artificial Intelligence Pulse Coupled Neural Network Algorithm in the Diagnosis and Treatment of Severe Sepsis Complicated with Acute Kidney Injury under Ultrasound Image. J. Healthc. Eng. 2021, 2021, 6761364. [Google Scholar] [CrossRef] [PubMed]
  6. Xu, L.; Wang, X.; Pu, P.; Li, S.; Shao, Y.; Li, Y. Ultrasonic Image Features under the Intelligent Algorithm in the Diagnosis of Severe Sepsis Complicated with Renal Injury. Comput. Math. Methods Med. 2022, 2022, 2310014, Retraction in: Comput. Math. Methods Med. 2023, 2023, 9843757. [Google Scholar] [CrossRef]
  7. Greenstein, Y.Y.; Guevarra, K. Point-of-Care Ultrasound in the Intensive Care Unit Applications, Limitations, and the Evolution of Clinical Practice. Clin. Chest Med. 2022, 43, 373–384. [Google Scholar] [CrossRef] [PubMed]
  8. Nhat, P.T.H.; Van Hao, N.; Tho, P.V.; Kerdegari, H.; Pisani, L.; Thu, L.N.M.; Phuong, L.T.; Duong, H.T.H.; Thuy, D.B.; McBride, A.; et al. Clinical benefit of AI-assisted lung ultrasound in a resource-limited intensive care unit. Crit. Care 2023, 27, 257. [Google Scholar] [CrossRef] [PubMed]
  9. Cheema, B.S.; Walter, J.; Narang, A.; Thomas, J.D. Artificial Intelligence-Enabled POCUS in the COVID-19 ICU: A New Spin on Cardiac Ultrasound. JACC Case Rep. 2021, 3, 258–263. [Google Scholar] [CrossRef] [PubMed]
  10. Zhai, S.; Wang, H.; Sun, L.; Zhang, B.; Huo, F.; Qiu, S.; Wu, X.; Ma, J.; Wu, Y.; Duan, J. Artificial intelligence (AI) versus expert: A comparison of left ventricular outflow tract velocity time integral (LVOT-VTI) assessment between ICU doctors and an AI tool. J. Appl. Clin. Med. Phys. 2022, 23, e13724. [Google Scholar] [CrossRef]
  11. Stewart, J.E.; Goudie, A.; Mukherjee, A.; Dwivedi, G. Artificial intelligence-enhanced echocardiography in the emergency department. Emerg. Med. Australas. 2021, 33, 1117–1120. [Google Scholar] [CrossRef]
  12. Yıldız Potter, İ.; Leo, M.M.; Vaziri, A.; Feldman, J.A. Automated detection and localization of pericardial effusion from point-of-care cardiac ultrasound examination. Med. Biol. Eng. Comput. 2023, 61, 1947–1959. [Google Scholar] [CrossRef] [PubMed]
  13. Nti, B.; Lehmann, A.S.; Haddad, A.; Kennedy, S.K.; Russell, F.M. Artificial Intelligence-Augmented Pediatric Lung POCUS: A Pilot Study of Novice Learners. J. Ultrasound Med. 2022, 41, 2965–2972. [Google Scholar] [CrossRef] [PubMed]
  14. Bottomley, T.; Gadsden, J.; West, S. The failed peripheral nerve block. BJA Educ. 2023, 23, 92–100. [Google Scholar] [CrossRef]
  15. Czym jest Deep Learning? Available online: https://thestory.is/pl/journal/co-to-jest-deep-learning/ (accessed on 1 January 2024).
  16. Gohar, E.; Herling, A.; Mazuz, M.; Tsaban, G.; Gat, T.; Kobal, S.; Fuchs, L. Artificial Intelligence (AI) versus POCUS Expert: A Validation Study of Three Automatic AI-Based, Real-Time, Hemodynamic Echocardiographic Assessment Tools. J. Clin. Med. 2023, 12, 1352. [Google Scholar] [CrossRef]
  17. Shaikh, F.; Kenny, J.E.; Awan, O.; Markovic, D.; Friedman, O.; He, T.; Singh, S.; Yan, P.; Qadir, N.; Barjaktarevic, I. Measuring the accuracy of cardiac output using POCUS: The introduction of artificial intelligence into routine care. Ultrasound J. 2022, 14, 47. [Google Scholar] [CrossRef]
  18. Shiran, C.; Barjaktarevic, I. VenueTM Family, Auto Tool for Measuring VTI. GE Healthcare, White Paper 2021. Available online: https://www.pocushub.net/data/venue-family_auto-vti_whitepaper_pocus__glob_jb02194xx.pdf (accessed on 1 January 2024).
  19. Walpot, J.; Vermeiren, G.; Al Mafragi, A.; Malbrain, L.N.G.M. Comprehensive assessment of the aortic valve in critically ill patients for the non-cardiologist. Part I: Aortic stenosis of the native valve. Anaesthesiol. Intensive Ther. 2021, 53, 37–54. [Google Scholar] [CrossRef]
  20. Point of Care. Available online: https://www.mindray.com/en/solutions/medical-imaging/point-of-care (accessed on 1 January 2024).
  21. Enhance Your Echo Lab Workflow. Available online: https://www.dia-analysis.com/lvivoseamless (accessed on 1 January 2024).
  22. Kosmos Artificial Intelligence. Available online: https://echonous.com/product/kosmos-ai/ (accessed on 1 January 2024).
  23. Butterfly Education. Available online: https://www.butterflynetwork.com/education (accessed on 1 January 2024).
  24. Blehar, D.J.; Resop, D.; Chin, B.; Dayno, M.; Gaspari, R. Inferior vena cava displacement during respirophasic ultrasound imaging. Crit. Ultrasound J. 2012, 4, 18. [Google Scholar] [CrossRef] [PubMed]
  25. Ruge, M.; Marhefka, G.D. IVC measurement for the noninvasive evaluation of central venous pressure. J. Echocardiogr. 2022, 20, 133–143. [Google Scholar] [CrossRef]
  26. Mitchell, C.; Rahko, P.S.; Blauwet, L.A.; Canaday, B.; Finstuen, J.A.; Foster, M.C.; Horton, K.; Ogunyankin, K.O.; Palma, R.A.; Velazquez, E.J. Guidelines for Performing a Comprehensive Transthoracic Echocardiographic Examination in Adults: Recommendations from the American Society of Echocardiography. J. Am. Soc. Echocardiogr. 2019, 32, 1–64. [Google Scholar] [CrossRef]
  27. Yildizdas, D.; Aslan, N. Ultrasonographic inferior vena cava collapsibility and distensibility indices for detecting the volume status of critically ill pediatric patients. J. Ultrason. 2020, 20, e205–e209. [Google Scholar] [CrossRef]
  28. Barbier, C.; Loubières, Y.; Schmit, C.; Hayon, J.; Ricôme, J.L.; Jardin, F.; Vieillard-Baron, A. Respiratory changes in inferior vena cava diameter are helpful in predicting fluid responsiveness in ventilated septic patients. Intensive Care Med. 2004, 30, 1740–1746. [Google Scholar] [CrossRef]
  29. Vignon, P. Evaluation of fluid responsiveness in ventilated septic patients: Back to venous return. Intensive Care Med. 2004, 30, 1699–1701. [Google Scholar] [CrossRef]
  30. Moretti, R.; Pizzi, B. Inferior Vena Cava Distensibility as a Predictor of Fluid Responsiveness in Patients with Subarachnoid Hemorrhage. Neurocritical. Care 2010, 13, 3–9. [Google Scholar] [CrossRef]
  31. Guiotto, G.; Masarone, M.; Paladino, F.; Ruggiero, E.; Scott, S.; Verde, S.; Schiraldi, F. Inferior vena cava collapsibility to guide fluid removal in slow continuous ultrafiltration: A pilot study. Intensive Care Med. 2010, 36, 692–696. [Google Scholar] [CrossRef] [PubMed]
  32. Duller, A.; Nickson, C. Airway Ultrasound. Available online: https://litfl.com/airway-ultrasound/ (accessed on 1 January 2024).
  33. Wang, H.; Shikano, K.; Nakajima, T.; Nomura, Y.; Nakaguchi, T. Peripheral Pulmonary Lesions Classification Using Endobronchial Ultrasonography Images Based on Bagging Ensemble Learning and Down-Sampling Technique. Appl. Sci. 2023, 13, 8403. [Google Scholar] [CrossRef]
  34. Tan, G.F.L.; Du, T.; Liu, J.S.; Chai, C.C.; Nyein, C.M.; Liu, A.Y.L. Automated lung ultrasound image assessment using artifcial intelligence to identify fuid overload in dialysis patients. BMC Nephrol. 2022, 23, 410. [Google Scholar] [CrossRef]
  35. Damodaran, S.; Kulkarni, A.V.; Gunaseelan, V.; Raj, V.; Kanchi, M. Automated versus manual B-lines counting, left ventricular outflow tract velocity time integral and inferior vena cava collapsibility index in COVID-19 patients. Indian J. Anaesth. 2022, 66, 368–374. [Google Scholar] [PubMed]
  36. El Khoury, D.; Pardo, E.; Cambriel, A.; Bonnet, F.; Pham, T.; Cholley, B.; Quesnel, C.; Verdonk, F. Gastric Cross-Sectional Area to Predict Gastric Intolerance in Critically Ill Patients: The Sono-ICU Prospective Observational Bicenter Study. Crit. Care Explor. 2023, 5, e0882. [Google Scholar] [CrossRef]
  37. NEW Clarius PAL HD3 Wireless Dual-Array for Whole-Body Ultrasound! Available online: https://clarius.com/pal-dual-array-ultrasound/ (accessed on 1 January 2024).
  38. Exo 2023 Survey Report. Unlocking Point-of-Care Ultrasound. Available online: https://www.exo.inc/article/2023-survey-report-unlocking-point-of-care-ultrasound (accessed on 1 January 2024).
Figure 1. Manual VTI measurement.
Figure 1. Manual VTI measurement.
Jpm 14 00286 g001
Figure 2. SmartVTI and a trending graph (Mindray TEX20; software version 01(01.07.00)).
Figure 2. SmartVTI and a trending graph (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g002
Figure 3. SmartEchoVue (Mindray TEX20; software version 01(01.07.00)).
Figure 3. SmartEchoVue (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g003
Figure 4. LVivo Seamless (Philips Epiq Elite, software version 9.05).
Figure 4. LVivo Seamless (Philips Epiq Elite, software version 9.05).
Jpm 14 00286 g004
Figure 5. AI Trio (EchoNous, Inc.), adapted from (source: https://echonous.com/product/kosmos-ai/) with permission publisher 2024 (accessed on 1 January 2024) [22].
Figure 5. AI Trio (EchoNous, Inc.), adapted from (source: https://echonous.com/product/kosmos-ai/) with permission publisher 2024 (accessed on 1 January 2024) [22].
Jpm 14 00286 g005
Figure 6. Butterfly ScanLab (Butterfly Network, Inc.), adapted from (https://www.butterflynetwork.com/education) with permission publisher 2024 (accessed on 1 January 2024) [23].
Figure 6. Butterfly ScanLab (Butterfly Network, Inc.), adapted from (https://www.butterflynetwork.com/education) with permission publisher 2024 (accessed on 1 January 2024) [23].
Jpm 14 00286 g006
Figure 7. EF measurement on 4CH and 2 CH (Mindray TEX20; software version 01(01.07.00)).
Figure 7. EF measurement on 4CH and 2 CH (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g007
Figure 8. Method for calculating EF using the Simpson method.
Figure 8. Method for calculating EF using the Simpson method.
Jpm 14 00286 g008
Figure 9. Automatic EF measurement (Mindray TEX20; software version 01(01.07.00)).
Figure 9. Automatic EF measurement (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g009
Figure 10. Imaging variability of the inferior vena cava depending on the respiratory phase.
Figure 10. Imaging variability of the inferior vena cava depending on the respiratory phase.
Jpm 14 00286 g010
Figure 11. Conventional IVC measurement.
Figure 11. Conventional IVC measurement.
Jpm 14 00286 g011
Figure 12. Auto-measurement of IVC using SmartIVC (Mindray TEX20; software version 01(01.07.00)).
Figure 12. Auto-measurement of IVC using SmartIVC (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g012
Figure 13. Zone annotation screen (Mindray TEX20; software version 01(01.07.00)).
Figure 13. Zone annotation screen (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g013
Figure 14. Smart B-line technology (Mindray TEX20; software version 01(01.07.00)).
Figure 14. Smart B-line technology (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g014
Figure 15. A scoring map and full review—Smart B-line (Mindray TEX20; software version 01(01.07.00)).
Figure 15. A scoring map and full review—Smart B-line (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g015
Figure 16. Automatic detection of A-line in lung ultrasound, adapted from (https://www.butterflynetwork.com/iq-plus) with permission publisher 2024 (accessed on 1 January 2024) [29].
Figure 16. Automatic detection of A-line in lung ultrasound, adapted from (https://www.butterflynetwork.com/iq-plus) with permission publisher 2024 (accessed on 1 January 2024) [29].
Jpm 14 00286 g016
Figure 17. Auto Gastric Antrum (Mindray TEX20; software version 01(01.07.00)).
Figure 17. Auto Gastric Antrum (Mindray TEX20; software version 01(01.07.00)).
Jpm 14 00286 g017
Figure 18. Clarius PAL HD3, adapted from (https://clarius.com/pal-dual-array-ultrasound) with permission publisher 2024 (accessed on 1 January 2024) [37].
Figure 18. Clarius PAL HD3, adapted from (https://clarius.com/pal-dual-array-ultrasound) with permission publisher 2024 (accessed on 1 January 2024) [37].
Jpm 14 00286 g018
Figure 19. Value added from AI, reproduced from (Exo 2023 Survey Report. Unlocking Point-of-Care Ultrasound) [38].
Figure 19. Value added from AI, reproduced from (Exo 2023 Survey Report. Unlocking Point-of-Care Ultrasound) [38].
Jpm 14 00286 g019
Table 1. Summary of selected technological and artificial intelligence in ultrasonography in the field of ICU.
Table 1. Summary of selected technological and artificial intelligence in ultrasonography in the field of ICU.
AI TechnologyManufacturerShort DescriptionRange of Ultrasound Exam
LvivoPhilipsautomatic measurement of stroke volume (SV) and cardiac output (CO)Ultrasonographic assessment of cardiac function
AutoVTIGE
US2.AIEchoNous
SmartVTIMindray
SmartEchoVueMindrayassessment of views Whole body
Us2.ai and AI TRO EchoNous
Butterfly ScanLab Butterfly Network
LVivo SeamlessPhilips
AutoEFMindraycalculation of ejection fraction (EF)Ultrasonographic assessment of cardiac function
RealTimeEFGE
US2.AIEchoNous
LvivoEFPhilips
SmartIVCMindrayautomatic measurement of the inferior vena cava (IVC)Internal Vena Cava examination
AutoIVCGE
Smart B-lineMindrayautomatic identification of the artifacts Lung examination
Butterfly ScanLab Butterfly Network
Auto B-lineGE
Auto Gastric Antrum Mindrayautomatic measurement of gastric contentsGastric examination
Clarius PAL HD3 Clarius Mobile Health1 probe for each examinationWhole body
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mika, S.; Gola, W.; Gil-Mika, M.; Wilk, M.; Misiolłek, H. Ultrasonographic Applications of Novel Technologies and Artificial Intelligence in Critically Ill Patients. J. Pers. Med. 2024, 14, 286. https://doi.org/10.3390/jpm14030286

AMA Style

Mika S, Gola W, Gil-Mika M, Wilk M, Misiolłek H. Ultrasonographic Applications of Novel Technologies and Artificial Intelligence in Critically Ill Patients. Journal of Personalized Medicine. 2024; 14(3):286. https://doi.org/10.3390/jpm14030286

Chicago/Turabian Style

Mika, Sławomir, Wojciech Gola, Monika Gil-Mika, Mateusz Wilk, and Hanna Misiolłek. 2024. "Ultrasonographic Applications of Novel Technologies and Artificial Intelligence in Critically Ill Patients" Journal of Personalized Medicine 14, no. 3: 286. https://doi.org/10.3390/jpm14030286

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop