Next Issue
Volume 9, February
Previous Issue
Volume 8, December
 
 

J. Imaging, Volume 9, Issue 1 (January 2023) – 19 articles

Cover Story (view full-size image): Seven centuries after the first Augmented Reality (AR) system, discussed as such first in this paper, three decades after the first series of Medical AR (MAR) technologies were published, and ten years after the deployment of the first MAR solutions into operating rooms, the exact definition, basic components, systematic design, and validation of MAR systems still lack a detailed discussion. This paper defines the basic components of any AR solution and extends them to exemplary MAR systems. We exemplify our framework with original MAR applications developed at the Chair for Computer Aided Medical Procedures and deployed into medical schools for teaching and into ORs for telemedicine and surgical guidance throughout the last few decades. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
13 pages, 4931 KiB  
Article
Picomolar Detection of Lead Ions (Pb2+) by Functionally Modified Fluorescent Carbon Quantum Dots from Watermelon Juice and Their Imaging in Cancer Cells
by Kundan Singh Rawat, Vikram Singh, Chandra Prakash Sharma, Akanksha Vyas, Priyanka Pandey, Jagriti Singh, Neeraj Mohan Gupta, Monika Sachdev and Atul Goel
J. Imaging 2023, 9(1), 19; https://doi.org/10.3390/jimaging9010019 - 16 Jan 2023
Cited by 6 | Viewed by 2748
Abstract
Water contamination due to the presence of lead is one of the leading causes of environmental and health hazards because of poor soil and groundwater waste management. Herein we report the synthesis of functionally modified luminescent carbon quantum dots (CQDs) obtained from watermelon [...] Read more.
Water contamination due to the presence of lead is one of the leading causes of environmental and health hazards because of poor soil and groundwater waste management. Herein we report the synthesis of functionally modified luminescent carbon quantum dots (CQDs) obtained from watermelon juice as potential nanomaterials for the detection of toxic Pb2+ ions in polluted water and cancer cells. By introducing surface passivating ligands such as ethanolamine (EA) and ethylenediamine (ED) in watermelon juice, watermelon-ethanolamine (WMEA)-CQDs and watermelon-ethylenediamine (WMED)-CQDs exhibited a remarkable ~10-fold and ~6-fold increase in fluorescence intensity with respect to non-doped WM-CQDs. The relative fluorescence quantum yields of WMEA-CQDs and WMED-CQDs were found to be 8% and 7%, respectively, in an aqueous medium. Among various functionally-modified CQDs, only WMED-CQDs showed high selectivity towards Pb2+ ions with a remarkably good limit of detection (LoD) of 190 pM, which is less than that of the permissible limit (72 nM) in drinking water. The functionally altered WMED-CQDs detected Pb2+ metal ions in polluted water and in a human cervical cancer cell line (HeLa), thus advocating new vistas for eco-friendly nanomaterials for their use as diagnostic tools in the environment and biomedical research areas. Full article
(This article belongs to the Special Issue Fluorescence Imaging and Analysis of Cellular System)
Show Figures

Figure 1

16 pages, 2230 KiB  
Review
Deepfakes Generation and Detection: A Short Survey
by Zahid Akhtar
J. Imaging 2023, 9(1), 18; https://doi.org/10.3390/jimaging9010018 - 13 Jan 2023
Cited by 14 | Viewed by 22816
Abstract
Advancements in deep learning techniques and the availability of free, large databases have made it possible, even for non-technical people, to either manipulate or generate realistic facial samples for both benign and malicious purposes. DeepFakes refer to face multimedia content, which has been [...] Read more.
Advancements in deep learning techniques and the availability of free, large databases have made it possible, even for non-technical people, to either manipulate or generate realistic facial samples for both benign and malicious purposes. DeepFakes refer to face multimedia content, which has been digitally altered or synthetically created using deep neural networks. The paper first outlines the readily available face editing apps and the vulnerability (or performance degradation) of face recognition systems under various face manipulations. Next, this survey presents an overview of the techniques and works that have been carried out in recent years for deepfake and face manipulations. Especially, four kinds of deepfake or face manipulations are reviewed, i.e., identity swap, face reenactment, attribute manipulation, and entire face synthesis. For each category, deepfake or face manipulation generation methods as well as those manipulation detection methods are detailed. Despite significant progress based on traditional and advanced computer vision, artificial intelligence, and physics, there is still a huge arms race surging up between attackers/offenders/adversaries (i.e., DeepFake generation methods) and defenders (i.e., DeepFake detection methods). Thus, open challenges and potential research directions are also discussed. This paper is expected to aid the readers in comprehending deepfake generation and detection mechanisms, together with open issues and future directions. Full article
Show Figures

Figure 1

19 pages, 21882 KiB  
Article
End-to-End Transcript Alignment of 17th Century Manuscripts: The Case of Moccia Code
by Giuseppe De Gregorio, Giuliana Capriolo and Angelo Marcelli
J. Imaging 2023, 9(1), 17; https://doi.org/10.3390/jimaging9010017 - 13 Jan 2023
Cited by 1 | Viewed by 1371
Abstract
The growth of digital libraries has yielded a large number of handwritten historical documents in the form of images, often accompanied by a digital transcription of the content. The ability to track the position of the words of the digital transcription in the [...] Read more.
The growth of digital libraries has yielded a large number of handwritten historical documents in the form of images, often accompanied by a digital transcription of the content. The ability to track the position of the words of the digital transcription in the images can be important both for the study of the document by humanities scholars and for further automatic processing. We propose a learning-free method for automatically aligning the transcription to the document image. The method receives as input the digital image of the document and the transcription of its content and aims at linking the transcription to the corresponding images within the page at the word level. The method comprises two main original contributions: a line-level segmentation algorithm capable of detecting text lines with curved baseline, and a text-to-image alignment algorithm capable of dealing with under- and over-segmentation errors at the word level. Experiments on pages from a 17th-century Italian manuscript have demonstrated that the line segmentation method allows one to segment 92% of the text line correctly. They also demonstrated that it achieves a correct alignment accuracy greater than 68%. Moreover, the performance achieved on widely used data sets compare favourably with the state of the art. Full article
Show Figures

Figure 1

16 pages, 6362 KiB  
Article
Stratigraphy of Fresco Paintings: A New Approach with Photoacoustic and SORS Imaging
by Francesca A. Pisu, Daniele Chiriu, Evgenia Klironomou, Giannis Zacharakis and George J. Tserevelakis
J. Imaging 2023, 9(1), 16; https://doi.org/10.3390/jimaging9010016 - 11 Jan 2023
Cited by 2 | Viewed by 1393
Abstract
Photoacoustic (PA) imaging is a novel, powerful diagnostic technique utilized in different research fields. In particular, during recent years it has found several applications in Cultural Heritage (CH) diagnostics. PA imaging can be realized in transmittance or epi-illumination (reflectance) modes, obtaining variable levels [...] Read more.
Photoacoustic (PA) imaging is a novel, powerful diagnostic technique utilized in different research fields. In particular, during recent years it has found several applications in Cultural Heritage (CH) diagnostics. PA imaging can be realized in transmittance or epi-illumination (reflectance) modes, obtaining variable levels of contrast and spatial resolution. In this work, we confirmed the applicability of the PA technique as a powerful tool for the imaging of one of the most challenging artwork objects, namely fresco wall paints, to obtain precise stratigraphic profiles in different layered fresco samples. In this regard, we studied some multi-layered fragments of the vault of San Giuseppe Church in Cagliari (1870 AD) and some mock-ups realized specifically to test the potentiality of this technique. Due to complex structures of the frescoes, we used the Spatially Off-set Raman Spectroscopy (SORS) technique to provide complementary information. The experimental results were in agreement for both techniques, even for the three-layered complex structure, and were confirmed with Scanning Electron Microscopy (SEM) analysis of cross-sections. The combined use of these two techniques proved useful to investigate detailed hidden information on the fresco samples. Full article
(This article belongs to the Section Document Analysis and Processing)
Show Figures

Figure 1

11 pages, 7241 KiB  
Article
New Neutron Imaging Facility NIFFLER at Very Low Power Reactor VR-1
by Jana Matouskova, Burkhard Schillinger and Lubomir Sklenka
J. Imaging 2023, 9(1), 15; https://doi.org/10.3390/jimaging9010015 - 10 Jan 2023
Cited by 3 | Viewed by 1986
Abstract
The paper describes the construction of the neutron imaging facility at the very low-power research reactor VR-1. The training reactor VR-1 is operated by the Czech Technical University in Prague, Czech Republic. It is mainly used for the education of students in the [...] Read more.
The paper describes the construction of the neutron imaging facility at the very low-power research reactor VR-1. The training reactor VR-1 is operated by the Czech Technical University in Prague, Czech Republic. It is mainly used for the education of students in the field of nuclear engineering as well as for the training of professionals. Neutron imaging is the new field of VR-1 reactor utilisation currently under development. Extremely low reactor power at the level of 100 W brought many challenges that were necessary to overcome to build and commission a sustainable neutron radiography facility. The paper describes the reactor’s neutron flux verification and the basic concept and design of the neutron imaging instrumentation. The first experimental results were mainly dedicated to testing the detection system for different radial beam port configurations, different L/D ratios, and different exposure times. Preliminary results of neutron radiography and tomography measurements at VR-1 clearly showed the potential of using neutron imaging in low-power reactors such as the VR-1 reactor. Full article
Show Figures

Figure 1

35 pages, 4625 KiB  
Article
Social Media Devices’ Influence on User Neck Pain during the COVID-19 Pandemic: Collaborating Vertebral-GLCM Extracted Features with a Decision Tree
by Bassam Al-Naami, Bashar E. A. Badr, Yahia Z. Rawash, Hamza Abu Owida, Roberto De Fazio and Paolo Visconti
J. Imaging 2023, 9(1), 14; https://doi.org/10.3390/jimaging9010014 - 8 Jan 2023
Cited by 3 | Viewed by 2099
Abstract
The prevalence of neck pain, a chronic musculoskeletal disease, has significantly increased due to the uncontrollable use of social media (SM) devices. The use of SM devices by younger generations increased enormously during the COVID-19 pandemic, being—in some cases—the only possibility for maintaining [...] Read more.
The prevalence of neck pain, a chronic musculoskeletal disease, has significantly increased due to the uncontrollable use of social media (SM) devices. The use of SM devices by younger generations increased enormously during the COVID-19 pandemic, being—in some cases—the only possibility for maintaining interpersonal, social, and friendship relationships. This study aimed to predict the occurrence of neck pain and its correlation with the intensive use of SM devices. It is based on nine quantitative parameters extracted from the retrospective X-ray images. The three parameters related to angle_1 (i.e., the angle between the global horizontal and the vector pointing from C7 vertebra to the occipito-cervical joint), angle_2 (i.e., the angle between the global horizontal and the vector pointing from C1 vertebra to the occipito-cervical joint), and the area between them were measured from the shape of the neck vertebrae, while the rest of the parameters were extracted from the images using the gray-level co-occurrence matrix (GLCM). In addition, the users’ ages and the duration of the SM usage (H.mean) were also considered. The decision tree (DT) machine-learning algorithm was employed to predict the abnormal cases (painful subjects) against the normal ones (no pain). The results showed that angle_1, area, and the image contrast significantly increased statistically with the time of SM-device usage, precisely in the range of 2 to 9 h. The DT showed a promising result demonstrated by classification accuracy and F1-scores of 94% and 0.95, respectively. Our findings confirmed that the objectively detected parameters, which elucidate the negative impacts of SM-device usage on neck pain, can be predicted by DT machine learning. Full article
(This article belongs to the Special Issue Intelligent Strategies for Medical Image Analysis)
Show Figures

Figure 1

21 pages, 5736 KiB  
Article
Fuzzy Model for the Automatic Recognition of Human Dendritic Cells
by Marwa Braiki, Kamal Nasreddine, Abdesslam Benzinou and Nolwenn Hymery
J. Imaging 2023, 9(1), 13; https://doi.org/10.3390/jimaging9010013 - 6 Jan 2023
Viewed by 1851
Abstract
Background and objective: Nowadays, foodborne illness is considered one of the most outgrowing diseases in the world, and studies show that its rate increases sharply each year. Foodborne illness is considered a public health problem which is caused by numerous factors, such as [...] Read more.
Background and objective: Nowadays, foodborne illness is considered one of the most outgrowing diseases in the world, and studies show that its rate increases sharply each year. Foodborne illness is considered a public health problem which is caused by numerous factors, such as food intoxications, allergies, intolerances, etc. Mycotoxin is one of the food contaminants which is caused by various species of molds (or fungi), which, in turn, causes intoxications that can be chronic or acute. Thus, even low concentrations of Mycotoxin have a severely harmful impact on human health. It is, therefore, necessary to develop an assessment tool for evaluating their impact on the immune response. Recently, researchers have approved a new method of investigation using human dendritic cells, yet the analysis of the geometric properties of these cells is still visual. Moreover, this type of analysis is subjective, time-consuming, and difficult to perform manually. In this paper, we address the automation of this evaluation using image-processing techniques. Methods: Automatic classification approaches of microscopic dendritic cell images are developed to provide a fast and objective evaluation. The first proposed classifier is based on support vector machines (SVM) and Fisher’s linear discriminant analysis (FLD) method. The FLD–SVM classifier does not provide satisfactory results due to the significant confusion between the inhibited cells on one hand, and the other two cell types (mature and immature) on the other hand. Then, another strategy was suggested to enhance dendritic cell recognition results that are emitted from microscopic images. This strategy is mainly based on fuzzy logic which allows us to consider the uncertainties and inaccuracies of the given data. Results: These proposed methods are tested on a real dataset consisting of 421 images of microscopic dendritic cells, where the fuzzy classification scheme efficiently improved the classification results by successfully classifying 96.77% of the dendritic cells. Conclusions: The fuzzy classification-based tools provide cell maturity and inhibition rates which help biologists evaluate severe health impacts caused by food contaminants. Full article
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)
Show Figures

Figure 1

14 pages, 1690 KiB  
Article
An Approach toward Automatic Specifics Diagnosis of Breast Cancer Based on an Immunohistochemical Image
by Oleh Berezsky, Oleh Pitsun, Grygoriy Melnyk, Tamara Datsko, Ivan Izonin and Bohdan Derysh
J. Imaging 2023, 9(1), 12; https://doi.org/10.3390/jimaging9010012 - 4 Jan 2023
Cited by 4 | Viewed by 1968
Abstract
The paper explored the problem of automatic diagnosis based on immunohistochemical image analysis. The issue of automated diagnosis is a preliminary and advisory statement for a diagnostician. The authors studied breast cancer histological and immunohistochemical images using the following biomarkers progesterone, estrogen, oncoprotein, [...] Read more.
The paper explored the problem of automatic diagnosis based on immunohistochemical image analysis. The issue of automated diagnosis is a preliminary and advisory statement for a diagnostician. The authors studied breast cancer histological and immunohistochemical images using the following biomarkers progesterone, estrogen, oncoprotein, and a cell proliferation biomarker. The authors developed a breast cancer diagnosis method based on immunohistochemical image analysis. The proposed method consists of algorithms for image preprocessing, segmentation, and the determination of informative indicators (relative area and intensity of cells) and an algorithm for determining the molecular genetic breast cancer subtype. An adaptive algorithm for image preprocessing was developed to improve the quality of the images. It includes median filtering and image brightness equalization techniques. In addition, the authors developed a software module part of the HIAMS software package based on the Java programming language and the OpenCV computer vision library. Four molecular genetic breast cancer subtypes could be identified using this solution: subtype Luminal A, subtype Luminal B, subtype HER2/neu amplified, and basalt-like subtype. The developed algorithm for the quantitative characteristics of the immunohistochemical images showed sufficient accuracy in determining the cancer subtype “Luminal A”. It was experimentally established that the relative area of the nuclei of cells covered with biomarkers of progesterone, estrogen, and oncoprotein was more than 85%. The given approach allows for automating and accelerating the process of diagnosis. Developed algorithms for calculating the quantitative characteristics of cells on immunohistochemical images can increase the accuracy of diagnosis. Full article
Show Figures

Figure 1

15 pages, 701 KiB  
Review
Visualization and Cybersecurity in the Metaverse: A Survey
by Yang-Wai Chow, Willy Susilo, Yannan Li, Nan Li and Chau Nguyen
J. Imaging 2023, 9(1), 11; https://doi.org/10.3390/jimaging9010011 - 31 Dec 2022
Cited by 14 | Viewed by 4939
Abstract
The popularity of the Metaverse has rapidly increased in recent years. However, despite the attention, investment, and promise of the Metaverse, there are various cybersecurity issues that must be addressed before the Metaverse can truly be adopted in practice for serious applications. The [...] Read more.
The popularity of the Metaverse has rapidly increased in recent years. However, despite the attention, investment, and promise of the Metaverse, there are various cybersecurity issues that must be addressed before the Metaverse can truly be adopted in practice for serious applications. The realization of the Metaverse is envisioned by many as requiring the use of visualization technologies such as Virtual Reality (VR) and Augmented Reality (AR). This visual aspect of the Metaverse will undoubtedly give rise to emerging cybersecurity threats that have not received much attention. As such, the purpose of this survey is to investigate cybersecurity threats faced by the Metaverse in relation to visualization technologies. Furthermore, this paper discusses existing work and open research directions on the development of countermeasures against such threats. As the Metaverse is a multidisciplinary topic, the intention of this work is to provide a background of the field to aid researchers in related areas. Full article
(This article belongs to the Special Issue Visualisation and Cybersecurity)
Show Figures

Figure 1

22 pages, 5342 KiB  
Article
A Feature Extraction Using Probabilistic Neural Network and BTFSC-Net Model with Deep Learning for Brain Tumor Classification
by Arun Singh Yadav, Surendra Kumar, Girija Rani Karetla, Juan Carlos Cotrina-Aliaga, José Luis Arias-Gonzáles, Vinod Kumar, Satyajee Srivastava, Reena Gupta, Sufyan Ibrahim, Rahul Paul, Nithesh Naik, Babita Singla and Nisha S. Tatkar
J. Imaging 2023, 9(1), 10; https://doi.org/10.3390/jimaging9010010 - 31 Dec 2022
Cited by 10 | Viewed by 2908
Abstract
Background and Objectives: Brain Tumor Fusion-based Segments and Classification-Non-enhancing tumor (BTFSC-Net) is a hybrid system for classifying brain tumors that combine medical image fusion, segmentation, feature extraction, and classification procedures. Materials and Methods: to reduce noise from medical images, the hybrid probabilistic wiener [...] Read more.
Background and Objectives: Brain Tumor Fusion-based Segments and Classification-Non-enhancing tumor (BTFSC-Net) is a hybrid system for classifying brain tumors that combine medical image fusion, segmentation, feature extraction, and classification procedures. Materials and Methods: to reduce noise from medical images, the hybrid probabilistic wiener filter (HPWF) is first applied as a preprocessing step. Then, to combine robust edge analysis (REA) properties in magnetic resonance imaging (MRI) and computed tomography (CT) medical images, a fusion network based on deep learning convolutional neural networks (DLCNN) is developed. Here, the brain images’ slopes and borders are detected using REA. To separate the sick region from the color image, adaptive fuzzy c-means integrated k-means (HFCMIK) clustering is then implemented. To extract hybrid features from the fused image, low-level features based on the redundant discrete wavelet transform (RDWT), empirical color features, and texture characteristics based on the gray-level cooccurrence matrix (GLCM) are also used. Finally, to distinguish between benign and malignant tumors, a deep learning probabilistic neural network (DLPNN) is deployed. Results: according to the findings, the suggested BTFSC-Net model performed better than more traditional preprocessing, fusion, segmentation, and classification techniques. Additionally, 99.21% segmentation accuracy and 99.46% classification accuracy were reached using the proposed BTFSC-Net model. Conclusions: earlier approaches have not performed as well as our presented method for image fusion, segmentation, feature extraction, classification operations, and brain tumor classification. These results illustrate that the designed approach performed more effectively in terms of enhanced quantitative evaluation with better accuracy as well as visual performance. Full article
(This article belongs to the Topic Medical Image Analysis)
Show Figures

Figure 1

3 pages, 166 KiB  
Editorial
Computer Vision and Robotics for Cultural Heritage: Theory and Applications
by Guillaume Caron, Olga Regina Pereira Bellon and Ilan Shimshoni
J. Imaging 2023, 9(1), 9; https://doi.org/10.3390/jimaging9010009 - 30 Dec 2022
Viewed by 1640
Abstract
Computer vision and robotics are more and more involved in cultural heritage [...] Full article
15 pages, 30832 KiB  
Article
Automatic Method for Vickers Hardness Estimation by Image Processing
by Jonatan D. Polanco, Carlos Jacanamejoy-Jamioy, Claudia L. Mambuscay, Jeferson F. Piamba and Manuel G. Forero
J. Imaging 2023, 9(1), 8; https://doi.org/10.3390/jimaging9010008 - 30 Dec 2022
Cited by 4 | Viewed by 2217
Abstract
Hardness is one of the most important mechanical properties of materials, since it is used to estimate their quality and to determine their suitability for a particular application. One method of determining quality is the Vickers hardness test, in which the resistance to [...] Read more.
Hardness is one of the most important mechanical properties of materials, since it is used to estimate their quality and to determine their suitability for a particular application. One method of determining quality is the Vickers hardness test, in which the resistance to plastic deformation at the surface of the material is measured after applying force with an indenter. The hardness is measured from the sample image, which is a tedious, time-consuming, and prone to human error procedure. Therefore, in this work, a new automatic method based on image processing techniques is proposed, allowing for obtaining results quickly and more accurately even with high irregularities in the indentation mark. For the development and validation of the method, a set of microscopy images of samples indented with applied forces of 5N and 10N on AISI D2 steel with and without quenching, tempering heat treatment and samples coated with titanium niobium nitride (TiNbN) was used. The proposed method was implemented as a plugin of the ImageJ program, allowing for obtaining reproducible Vickers hardness results in an average time of 2.05 seconds with an accuracy of 98.3% and a maximum error of 4.5% with respect to the values obtained manually, used as a golden standard. Full article
(This article belongs to the Topic Computer Vision and Image Processing)
Show Figures

Figure 1

21 pages, 6798 KiB  
Article
The Optimization of the Light-Source Spectrum Utilizing Neural Networks for Detecting Oral Lesions
by Kenichi Ito, Hiroshi Higashi, Ari Hietanen, Pauli Fält, Kyoko Hine, Markku Hauta-Kasari and Shigeki Nakauchi
J. Imaging 2023, 9(1), 7; https://doi.org/10.3390/jimaging9010007 - 29 Dec 2022
Cited by 2 | Viewed by 1909
Abstract
Any change in the light-source spectrum modifies the color information of an object. The spectral distribution of the light source can be optimized to enhance specific details of the obtained images; thus, using information-enhanced images is expected to improve the image recognition performance [...] Read more.
Any change in the light-source spectrum modifies the color information of an object. The spectral distribution of the light source can be optimized to enhance specific details of the obtained images; thus, using information-enhanced images is expected to improve the image recognition performance via machine vision. However, no studies have applied light spectrum optimization to reduce the training loss in modern machine vision using deep learning. Therefore, we propose a method for optimizing the light-source spectrum to reduce the training loss using neural networks. A two-class classification of one-vs-rest among the classes, including enamel as a healthy condition and dental lesions, was performed to validate the proposed method. The proposed convolutional neural network-based model, which accepts a 5 × 5 small patch image, was compared with an alternating optimization scheme using a linear-support vector machine that optimizes classification weights and lighting weights separately. Furthermore, it was compared with the proposed neural network-based algorithm, which inputs a pixel and consists of fully connected layers. The results of the five-fold cross-validation revealed that, compared to the previous method, the proposed method improved the F1-score and was superior to the models that were using the immutable standard illuminant D65. Full article
Show Figures

Figure 1

13 pages, 1660 KiB  
Article
CAL-Tutor: A HoloLens 2 Application for Training in Obstetric Sonography and User Motion Data Recording
by Manuel Birlo, Philip J. Eddie Edwards, Soojeong Yoo, Brian Dromey, Francisco Vasconcelos, Matthew J. Clarkson and Danail Stoyanov
J. Imaging 2023, 9(1), 6; https://doi.org/10.3390/jimaging9010006 - 29 Dec 2022
Cited by 2 | Viewed by 2175
Abstract
Obstetric ultrasound (US) training teaches the relationship between foetal anatomy and the viewed US slice to enable navigation to standardised anatomical planes (head, abdomen and femur) where diagnostic measurements are taken. This process is difficult to learn, and results in considerable inter-operator variability. [...] Read more.
Obstetric ultrasound (US) training teaches the relationship between foetal anatomy and the viewed US slice to enable navigation to standardised anatomical planes (head, abdomen and femur) where diagnostic measurements are taken. This process is difficult to learn, and results in considerable inter-operator variability. We propose the CAL-Tutor system for US training based on a US scanner and phantom, where a model of both the baby and the US slice are displayed to the trainee in its physical location using the HoloLens 2. The intention is that AR guidance will shorten the learning curve for US trainees and improve spatial awareness. In addition to the AR guidance, we also record many data streams to assess user motion and the learning process. The HoloLens 2 provides eye gaze, head and hand position, ARToolkit and NDI Aurora tracking gives the US probe positions and an external camera records the overall scene. These data can provide a rich source for further analysis, such as distinguishing expert from novice motion. We have demonstrated the system in a sample of engineers. Feedback suggests that the system helps novice users navigate the US probe to the standard plane. The data capture is successful and initial data visualisations show that meaningful information about user behaviour can be captured. Initial feedback is encouraging and shows improved user assessment where AR guidance is provided. Full article
Show Figures

Figure 1

12 pages, 4431 KiB  
Article
A Parallax Image Mosaic Method for Low Altitude Aerial Photography with Artifact and Distortion Suppression
by Jing Xu, Dandan Zhao, Zhengwei Ren, Feiran Fu, Yuxin Sun and Ming Fang
J. Imaging 2023, 9(1), 5; https://doi.org/10.3390/jimaging9010005 - 25 Dec 2022
Viewed by 1543
Abstract
In this paper, we propose an aerial images stitching method based on an as-projective-as-possible (APAP) algorithm, aiming at the problem artifacts, distortions, or stitching failure due to fewer feature points for multispectral aerial image with certain parallax. Our method incorporates accelerated nonlinear diffusion [...] Read more.
In this paper, we propose an aerial images stitching method based on an as-projective-as-possible (APAP) algorithm, aiming at the problem artifacts, distortions, or stitching failure due to fewer feature points for multispectral aerial image with certain parallax. Our method incorporates accelerated nonlinear diffusion algorithm (AKAZE) into APAP algorithm. First, we use the fast and stable AKAZE to extract the feature points of aerial images, and then, based on the registration model of the APAP algorithm, we add line protection constraints, global similarity constraints, and local similarity constraints to protect the image structure information, to produce a panorama. Experimental results on several datasets demonstrate that proposed method is effective when dealing with multispectral aerial images. Our method can suppress artifacts, distortions, and reduce incomplete splicing. Compared with state-of-the-art image stitching methods, including APAP and adaptive as-natural-as-possible image stitching (AANAP), and two of the most popular UAV image stitching tools, Pix4D and OpenDroneMap (ODM), our method achieves them both quantitatively and qualitatively. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

21 pages, 26921 KiB  
Article
Medical Augmented Reality: Definition, Principle Components, Domain Modeling, and Design-Development-Validation Process
by Nassir Navab, Alejandro Martin-Gomez, Matthias Seibold, Michael Sommersperger, Tianyu Song, Alexander Winkler, Kevin Yu and Ulrich Eck
J. Imaging 2023, 9(1), 4; https://doi.org/10.3390/jimaging9010004 - 23 Dec 2022
Cited by 7 | Viewed by 4411
Abstract
Three decades after the first set of work on Medical Augmented Reality (MAR) was presented to the international community, and ten years after the deployment of the first MAR solutions into operating rooms, its exact definition, basic components, systematic design, and validation still [...] Read more.
Three decades after the first set of work on Medical Augmented Reality (MAR) was presented to the international community, and ten years after the deployment of the first MAR solutions into operating rooms, its exact definition, basic components, systematic design, and validation still lack a detailed discussion. This paper defines the basic components of any Augmented Reality (AR) solution and extends them to exemplary Medical Augmented Reality Systems (MARS). We use some of the original MARS applications developed at the Chair for Computer Aided Medical Procedures and deployed into medical schools for teaching anatomy and into operating rooms for telemedicine and surgical guidance throughout the last decades to identify the corresponding basic components. In this regard, the paper is not discussing all past or existing solutions but only aims at defining the principle components and discussing the particular domain modeling for MAR and its design-development-validation process, and providing exemplary cases through the past in-house developments of such solutions. Full article
Show Figures

Figure 1

11 pages, 632 KiB  
Article
Auguring Fake Face Images Using Dual Input Convolution Neural Network
by Mohan Bhandari, Arjun Neupane, Saurav Mallik, Loveleen Gaur and Hong Qin
J. Imaging 2023, 9(1), 3; https://doi.org/10.3390/jimaging9010003 - 21 Dec 2022
Cited by 10 | Viewed by 2953
Abstract
Deepfake technology uses auto-encoders and generative adversarial networks to replace or artificially construct fine-tuned faces, emotions, and sounds. Although there have been significant advancements in the identification of particular fake images, a reliable counterfeit face detector is still lacking, making it difficult to [...] Read more.
Deepfake technology uses auto-encoders and generative adversarial networks to replace or artificially construct fine-tuned faces, emotions, and sounds. Although there have been significant advancements in the identification of particular fake images, a reliable counterfeit face detector is still lacking, making it difficult to identify fake photos in situations with further compression, blurring, scaling, etc. Deep learning models resolve the research gap to correctly recognize phony images, whose objectionable content might encourage fraudulent activity and cause major problems. To reduce the gap and enlarge the fields of view of the network, we propose a dual input convolutional neural network (DICNN) model with ten-fold cross validation with an average training accuracy of 99.36 ± 0.62, a test accuracy of 99.08 ± 0.64, and a validation accuracy of 99.30 ± 0.94. Additionally, we used ’SHapley Additive exPlanations (SHAP) ’ as explainable AI (XAI) Shapely values to explain the results and interoperability visually by imposing the model into SHAP. The proposed model holds significant importance for being accepted by forensics and security experts because of its distinctive features and considerably higher accuracy than state-of-the-art methods. Full article
Show Figures

Figure 1

15 pages, 8796 KiB  
Article
Attenuation Correction Using Template PET Registration for Brain PET: A Proof-of-Concept Study
by Markus Jehl, Ekaterina Mikhaylova, Valerie Treyer, Marlena Hofbauer, Martin Hüllner, Philipp A. Kaufmann, Alfred Buck, Geoff Warnock, Viet Dao, Charalampos Tsoumpas, Daniel Deidda, Kris Thielemans, Max Ludwig Ahnen and Jannis Fischer
J. Imaging 2023, 9(1), 2; https://doi.org/10.3390/jimaging9010002 - 21 Dec 2022
Cited by 3 | Viewed by 1990
Abstract
NeuroLF is a dedicated brain PET system with an octagonal prism shape housed in a scanner head that can be positioned around a patient’s head. Because it does not have MR or CT capabilities, attenuation correction based on an estimation of the attenuation [...] Read more.
NeuroLF is a dedicated brain PET system with an octagonal prism shape housed in a scanner head that can be positioned around a patient’s head. Because it does not have MR or CT capabilities, attenuation correction based on an estimation of the attenuation map is a crucial feature. In this article, we demonstrate this method on [18F]FDG PET brain scans performed with a low-resolution proof of concept prototype of NeuroLF called BPET. We perform an affine registration of a template PET scan to the uncorrected emission image, and then apply the resulting transform to the corresponding template attenuation map. Using a whole-body PET/CT system as reference, we quantitively show that this method yields comparable image quality (0.893 average correlation to reference scan) to using the reference µ-map as obtained from the CT scan of the imaged patient (0.908 average correlation). We conclude from this initial study that attenuation correction using template registration instead of a patient CT delivers similar results and is an option for patients undergoing brain PET. Full article
Show Figures

Figure 1

31 pages, 7351 KiB  
Review
A Survey on Deep Learning in COVID-19 Diagnosis
by Xue Han, Zuojin Hu, Shuihua Wang and Yudong Zhang
J. Imaging 2023, 9(1), 1; https://doi.org/10.3390/jimaging9010001 - 20 Dec 2022
Cited by 10 | Viewed by 3178
Abstract
According to the World Health Organization statistics, as of 25 October 2022, there have been 625,248,843 confirmed cases of COVID-19, including 65,622,281 deaths worldwide. The spread and severity of COVID-19 are alarming. The economy and life of countries worldwide have been greatly affected. [...] Read more.
According to the World Health Organization statistics, as of 25 October 2022, there have been 625,248,843 confirmed cases of COVID-19, including 65,622,281 deaths worldwide. The spread and severity of COVID-19 are alarming. The economy and life of countries worldwide have been greatly affected. The rapid and accurate diagnosis of COVID-19 directly affects the spread of the virus and the degree of harm. Currently, the classification of chest X-ray or CT images based on artificial intelligence is an important method for COVID-19 diagnosis. It can assist doctors in making judgments and reduce the misdiagnosis rate. The convolutional neural network (CNN) is very popular in computer vision applications, such as applied to biological image segmentation, traffic sign recognition, face recognition, and other fields. It is one of the most widely used machine learning methods. This paper mainly introduces the latest deep learning methods and techniques for diagnosing COVID-19 using chest X-ray or CT images based on the convolutional neural network. It reviews the technology of CNN at various stages, such as rectified linear units, batch normalization, data augmentation, dropout, and so on. Several well-performing network architectures are explained in detail, such as AlexNet, ResNet, DenseNet, VGG, GoogleNet, etc. We analyzed and discussed the existing CNN automatic COVID-19 diagnosis systems from sensitivity, accuracy, precision, specificity, and F1 score. The systems use chest X-ray or CT images as datasets. Overall, CNN has essential value in COVID-19 diagnosis. All of them have good performance in the existing experiments. If expanding the datasets, adding GPU acceleration and data preprocessing techniques, and expanding the types of medical images, the performance of CNN will be further improved. This paper wishes to make contributions to future research. Full article
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop