Next Issue
Volume 8, May
Previous Issue
Volume 8, March
 
 

J. Imaging, Volume 8, Issue 4 (April 2022) – 36 articles

Cover Story (view full-size image): Modern deep neural networks are well known to be weak in the face of unknown data instances. Avoiding false predictions by identifying substantially different data from what has been seen during training remains a challenge. Although it is inevitable for continual-learning systems to encounter unseen concepts, the corresponding literature primarily focuses on preventing the catastrophic forgetting of learned representations. We bridge this gap by introducing the open variational auto-encoder (OpenVAE). OpenVAE unifies the detection of unseen, unknown, out-of-distribution data and the preservation of already acquired knowledge in continual training for robust application. View this paper.
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
15 pages, 17786 KiB  
Article
HISFCOS: Half-Inverted Stage Block for Efficient Object Detection Based on Deep Learning
by Beomyeon Hwang, Sanghun Lee and Seunghyun Lee
J. Imaging 2022, 8(4), 117; https://doi.org/10.3390/jimaging8040117 - 17 Apr 2022
Cited by 1 | Viewed by 1974
Abstract
Recent advances in object detection play a key role in various industrial applications. However, a fully convolutional one-stage detector (FCOS), a conventional object detection method, has low detection accuracy given the calculation cost. Thus, in this study, we propose a half-inverted stage FCOS [...] Read more.
Recent advances in object detection play a key role in various industrial applications. However, a fully convolutional one-stage detector (FCOS), a conventional object detection method, has low detection accuracy given the calculation cost. Thus, in this study, we propose a half-inverted stage FCOS (HISFCOS) with improved detection accuracy at a computational cost comparable to FCOS based on the proposed half inverted stage (HIS) block. First, FCOS has low detection accuracy owing to low-level information loss. Therefore, an HIS block that minimizes feature loss by extracting spatial and channel information in parallel is proposed. Second, detection accuracy was improved by reconstructing the feature pyramid on the basis of the proposed block and improving the low-level information. Lastly, the improved detection head structure reduced the computational cost and amount compared to the conventional method. Through experiments, the proposed method defined the optimal HISFCOS parameters and evaluated several datasets for fair comparison. The HISFCOS was trained and evaluated using the PASCAL VOC and MSCOCO2017 datasets. Additionally, the average precision (AP) was used as an evaluation index to quantitatively evaluate detection performance. As a result of the experiment, the parameters were increased by 0.5 M compared to the conventional method, but the detection accuracy was improved by 3.0 AP and 1.5 AP in the PASCAL VOC and MSCOCO datasets, respectively. in addition, an ablation study was conducted, and the results for the proposed block and detection head were analyzed. Full article
Show Figures

Figure 1

18 pages, 11405 KiB  
Article
A Comparison of Dense and Sparse Optical Flow Techniques for Low-Resolution Aerial Thermal Imagery
by Tran Xuan Bach Nguyen, Kent Rosser and Javaan Chahl
J. Imaging 2022, 8(4), 116; https://doi.org/10.3390/jimaging8040116 - 16 Apr 2022
Cited by 5 | Viewed by 2645
Abstract
It is necessary to establish the relative performance of established optical flow approaches in airborne scenarios with thermal cameras. This study investigated the performance of a dense optical flow algorithm on 14 bit radiometric images of the ground. While sparse techniques that rely [...] Read more.
It is necessary to establish the relative performance of established optical flow approaches in airborne scenarios with thermal cameras. This study investigated the performance of a dense optical flow algorithm on 14 bit radiometric images of the ground. While sparse techniques that rely on feature matching techniques perform very well with airborne thermal data in high-contrast thermal conditions, these techniques suffer in low-contrast scenes, where there are fewer detectable and distinct features in the image. On the other hand, some dense optical flow algorithms are highly amenable to parallel processing approaches compared to those that rely on tracking and feature detection. A Long-Wave Infrared (LWIR) micro-sensor and a PX4Flow optical sensor were mounted looking downwards on a drone. We compared the optical flow signals of a representative dense optical flow technique, the Image Interpolation Algorithm (I2A), to the Lucas–Kanade (LK) algorithm in OpenCV and the visible light optical flow results from the PX4Flow in both X and Y displacements. The I2A to LK was found to be generally comparable in performance and better in cold-soaked environments while suffering from the aperture problem in some scenes. Full article
Show Figures

Figure 1

22 pages, 5894 KiB  
Article
Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces
by Hicham Talaoubrid, Marina Vert, Khizar Hayat and Baptiste Magnier
J. Imaging 2022, 8(4), 115; https://doi.org/10.3390/jimaging8040115 - 16 Apr 2022
Cited by 4 | Viewed by 2088
Abstract
The purpose of this paper is to find the best way to track human subjects in fisheye images by considering the most common similarity measures in the function of various color spaces as well as the HOG. To this end, we have relied [...] Read more.
The purpose of this paper is to find the best way to track human subjects in fisheye images by considering the most common similarity measures in the function of various color spaces as well as the HOG. To this end, we have relied on videos taken by a fisheye camera wherein multiple human subjects were recorded walking simultaneously, in random directions. Using an existing deep-learning method for the detection of persons in fisheye images, bounding boxes are extracted each containing information related to a single person. Consequently, each bounding box can be described by color features, usually color histograms; with the HOG relying on object shapes and contours. These descriptors do not inform the same features and they need to be evaluated in the context of tracking in top-view fisheye images. With this in perspective, a distance is computed to compare similarities between the detected bounding boxes of two consecutive frames. To do so, we are proposing a rate function (S) in order to compare and evaluate together the six different color spaces and six distances, and with the HOG. This function links inter-distance (i.e., the distance between the images of the same person throughout the frames of the video) with intra-distance (i.e., the distance between images of different people throughout the frames). It enables ascertaining a given feature descriptor (color or HOG) mapped to a corresponding similarity function and hence deciding the most reliable one to compute the similarity or the difference between two segmented persons. All these comparisons lead to some interesting results, as explained in the later part of the article. Full article
(This article belongs to the Special Issue Edge Detection Evaluation)
Show Figures

Figure 1

18 pages, 477 KiB  
Article
Resources and Power Efficient FPGA Accelerators for Real-Time Image Classification
by Angelos Kyriakos, Elissaios-Alexios Papatheofanous, Charalampos Bezaitis and Dionysios Reisis
J. Imaging 2022, 8(4), 114; https://doi.org/10.3390/jimaging8040114 - 15 Apr 2022
Cited by 5 | Viewed by 2505
Abstract
A plethora of image and video-related applications involve complex processes that impose the need for hardware accelerators to achieve real-time performance. Among these, notable applications include the Machine Learning (ML) tasks using Convolutional Neural Networks (CNNs) that detect objects in image frames. Aiming [...] Read more.
A plethora of image and video-related applications involve complex processes that impose the need for hardware accelerators to achieve real-time performance. Among these, notable applications include the Machine Learning (ML) tasks using Convolutional Neural Networks (CNNs) that detect objects in image frames. Aiming at contributing to the CNN accelerator solutions, the current paper focuses on the design of Field-Programmable Gate Arrays (FPGAs) for CNNs of limited feature space to improve performance, power consumption and resource utilization. The proposed design approach targets the designs that can utilize the logic and memory resources of a single FPGA device and benefit mainly the edge, mobile and on-board satellite (OBC) computing; especially their image-processing- related applications. This work exploits the proposed approach to develop an FPGA accelerator for vessel detection on a Xilinx Virtex 7 XC7VX485T FPGA device (Advanced Micro Devices, Inc, Santa Clara, CA, USA). The resulting architecture operates on RGB images of size 80×80 or sliding windows; it is trained for the “Ships in Satellite Imagery” and by achieving frequency 270 MHz, completing the inference in 0.687 ms and consuming 5 watts, it validates the approach. Full article
(This article belongs to the Special Issue Image Processing Using FPGAs 2021)
Show Figures

Figure 1

22 pages, 12387 KiB  
Article
Fuzzy Information Discrimination Measures and Their Application to Low Dimensional Embedding Construction in the UMAP Algorithm
by Liliya A. Demidova and Artyom V. Gorchakov
J. Imaging 2022, 8(4), 113; https://doi.org/10.3390/jimaging8040113 - 15 Apr 2022
Cited by 7 | Viewed by 2979
Abstract
Dimensionality reduction techniques are often used by researchers in order to make high dimensional data easier to interpret visually, as data visualization is only possible in low dimensional spaces. Recent research in nonlinear dimensionality reduction introduced many effective algorithms, including t-distributed stochastic neighbor [...] Read more.
Dimensionality reduction techniques are often used by researchers in order to make high dimensional data easier to interpret visually, as data visualization is only possible in low dimensional spaces. Recent research in nonlinear dimensionality reduction introduced many effective algorithms, including t-distributed stochastic neighbor embedding (t-SNE), uniform manifold approximation and projection (UMAP), dimensionality reduction technique based on triplet constraints (TriMAP), and pairwise controlled manifold approximation (PaCMAP), aimed to preserve both the local and global structure of high dimensional data while reducing the dimensionality. The UMAP algorithm has found its application in bioinformatics, genetics, genomics, and has been widely used to improve the accuracy of other machine learning algorithms. In this research, we compare the performance of different fuzzy information discrimination measures used as loss functions in the UMAP algorithm while constructing low dimensional embeddings. In order to achieve this, we derive the gradients of the considered losses analytically and employ the Adam algorithm during the loss function optimization process. From the conducted experimental studies we conclude that the use of either the logarithmic fuzzy cross entropy loss without reduced repulsion or the symmetric logarithmic fuzzy cross entropy loss with sufficiently large neighbor count leads to better global structure preservation of the original multidimensional data when compared to the loss function used in the original UMAP algorithm implementation. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

12 pages, 850 KiB  
Review
Spectral Photon-Counting Computed Tomography: A Review on Technical Principles and Clinical Applications
by Mario Tortora, Laura Gemini, Imma D’Iglio, Lorenzo Ugga, Gaia Spadarella and Renato Cuocolo
J. Imaging 2022, 8(4), 112; https://doi.org/10.3390/jimaging8040112 - 15 Apr 2022
Cited by 38 | Viewed by 5240
Abstract
Photon-counting computed tomography (CT) is a technology that has attracted increasing interest in recent years since, thanks to new-generation detectors, it holds the promise to radically change the clinical use of CT imaging. Photon-counting detectors overcome the major limitations of conventional CT detectors [...] Read more.
Photon-counting computed tomography (CT) is a technology that has attracted increasing interest in recent years since, thanks to new-generation detectors, it holds the promise to radically change the clinical use of CT imaging. Photon-counting detectors overcome the major limitations of conventional CT detectors by providing very high spatial resolution without electronic noise, providing a higher contrast-to-noise ratio, and optimizing spectral images. Additionally, photon-counting CT can lead to reduced radiation exposure, reconstruction of higher spatial resolution images, reduction of image artifacts, optimization of the use of contrast agents, and create new opportunities for quantitative imaging. The aim of this review is to briefly explain the technical principles of photon-counting CT and, more extensively, the potential clinical applications of this technology. Full article
(This article belongs to the Special Issue Intelligent Strategies for Medical Image Analysis)
Show Figures

Figure 1

8 pages, 1045 KiB  
Article
Reliability of OMERACT Scoring System in Ultra-High Frequency Ultrasonography of Minor Salivary Glands: Inter-Rater Agreement Study
by Rossana Izzetti, Giovanni Fulvio, Marco Nisi, Stefano Gennai and Filippo Graziani
J. Imaging 2022, 8(4), 111; https://doi.org/10.3390/jimaging8040111 - 15 Apr 2022
Cited by 5 | Viewed by 2131
Abstract
Minor salivary gland ultra-high frequency ultrasonography (UHFUS) has recently been introduced for the evaluation of patients with suspected primary Sjögren’s Syndrome (pSS). At present, ultrasonographic assessment of major salivary glands is performed using the Outcome Measures in Rheumatology (OMERACT) scoring system. Previous reports [...] Read more.
Minor salivary gland ultra-high frequency ultrasonography (UHFUS) has recently been introduced for the evaluation of patients with suspected primary Sjögren’s Syndrome (pSS). At present, ultrasonographic assessment of major salivary glands is performed using the Outcome Measures in Rheumatology (OMERACT) scoring system. Previous reports have explored the possibility of applying the OMERACT scoring system to minor salivary glands UHFUS, with promising results. The aim of this study was to test the inter-reader concordance in the assignment of the OMERACT score to minor salivary gland UHFUS. The study was conducted on 170 minor salivary glands UHFUS scans of patients with suspected pSS. Three independent readers performed UHFUS image evaluation. Intraclass correlation coefficient (ICC) was employed to assess inter-reader reliability. Bland and Altman analysis was employed to test the agreement with a gold standard examiner. ICC values > 0.9 were found for scores 0 and 1, while score 2 and score 3 presented ICCs of 0.873 and 0.785, respectively. The measurements performed by the three examiners were in agreement with the gold standard examiner. According to these results, UHFUS interpretation showed good inter-observer reliability, suggesting that OMERACT score can be effectively used for the evaluation of glandular alterations, even for minor salivary glands. Full article
(This article belongs to the Special Issue New Frontiers of Advanced Imaging in Dentistry)
Show Figures

Figure 1

21 pages, 5792 KiB  
Article
Salient Object Detection by LTP Texture Characterization on Opposing Color Pairs under SLICO Superpixel Constraint
by Didier Ndayikengurukiye and Max Mignotte
J. Imaging 2022, 8(4), 110; https://doi.org/10.3390/jimaging8040110 - 13 Apr 2022
Cited by 5 | Viewed by 3204
Abstract
The effortless detection of salient objects by humans has been the subject of research in several fields, including computer vision, as it has many applications. However, salient object detection remains a challenge for many computer models dealing with color and textured images. Most [...] Read more.
The effortless detection of salient objects by humans has been the subject of research in several fields, including computer vision, as it has many applications. However, salient object detection remains a challenge for many computer models dealing with color and textured images. Most of them process color and texture separately and therefore implicitly consider them as independent features which is not the case in reality. Herein, we propose a novel and efficient strategy, through a simple model, almost without internal parameters, which generates a robust saliency map for a natural image. This strategy consists of integrating color information into local textural patterns to characterize a color micro-texture. It is the simple, yet powerful LTP (Local Ternary Patterns) texture descriptor applied to opposing color pairs of a color space that allows us to achieve this end. Each color micro-texture is represented by a vector whose components are from a superpixel obtained by the SLICO (Simple Linear Iterative Clustering with zero parameter) algorithm, which is simple, fast and exhibits state-of-the-art boundary adherence. The degree of dissimilarity between each pair of color micro-textures is computed by the FastMap method, a fast version of MDS (Multi-dimensional Scaling) that considers the color micro-textures’ non-linearity while preserving their distances. These degrees of dissimilarity give us an intermediate saliency map for each RGB (Red–Green–Blue), HSL (Hue–Saturation–Luminance), LUV (L for luminance, U and V represent chromaticity values) and CMY (Cyan–Magenta–Yellow) color space. The final saliency map is their combination to take advantage of the strength of each of them. The MAE (Mean Absolute Error), MSE (Mean Squared Error) and Fβ measures of our saliency maps, on the five most used datasets show that our model outperformed several state-of-the-art models. Being simple and efficient, our model could be combined with classic models using color contrast for a better performance. Full article
(This article belongs to the Special Issue Advances in Color Imaging)
Show Figures

Figure 1

20 pages, 4325 KiB  
Article
Tracking Highly Similar Rat Instances under Heavy Occlusions: An Unsupervised Deep Generative Pipeline
by Anna Gelencsér-Horváth, László Kopácsi, Viktor Varga, Dávid Keller, Árpád Dobolyi, Kristóf Karacs and András Lőrincz
J. Imaging 2022, 8(4), 109; https://doi.org/10.3390/jimaging8040109 - 13 Apr 2022
Cited by 1 | Viewed by 4229
Abstract
Identity tracking and instance segmentation are crucial in several areas of biological research. Behavior analysis of individuals in groups of similar animals is a task that emerges frequently in agriculture or pharmaceutical studies, among others. Automated annotation of many hours of surveillance videos [...] Read more.
Identity tracking and instance segmentation are crucial in several areas of biological research. Behavior analysis of individuals in groups of similar animals is a task that emerges frequently in agriculture or pharmaceutical studies, among others. Automated annotation of many hours of surveillance videos can facilitate a large number of biological studies/experiments, which otherwise would not be feasible. Solutions based on machine learning generally perform well in tracking and instance segmentation; however, in the case of identical, unmarked instances (e.g., white rats or mice), even state-of-the-art approaches can frequently fail. We propose a pipeline of deep generative models for identity tracking and instance segmentation of highly similar instances, which, in contrast to most region-based approaches, exploits edge information and consequently helps to resolve ambiguity in heavily occluded cases. Our method is trained by synthetic data generation techniques, not requiring prior human annotation. We show that our approach greatly outperforms other state-of-the-art unsupervised methods in identity tracking and instance segmentation of unmarked rats in real-world laboratory video recordings. Full article
(This article belongs to the Special Issue Unsupervised Deep Learning and Its Applications in Imaging Processing)
Show Figures

Figure 1

29 pages, 23343 KiB  
Article
Multi-Stage Platform for (Semi-)Automatic Planning in Reconstructive Orthopedic Surgery
by Florian Kordon, Andreas Maier, Benedict Swartman, Maxim Privalov, Jan Siad El Barbari and Holger Kunze
J. Imaging 2022, 8(4), 108; https://doi.org/10.3390/jimaging8040108 - 12 Apr 2022
Cited by 3 | Viewed by 3088
Abstract
Intricate lesions of the musculoskeletal system require reconstructive orthopedic surgery to restore the correct biomechanics. Careful pre-operative planning of the surgical steps on 2D image data is an essential tool to increase the precision and safety of these operations. However, the plan’s effectiveness [...] Read more.
Intricate lesions of the musculoskeletal system require reconstructive orthopedic surgery to restore the correct biomechanics. Careful pre-operative planning of the surgical steps on 2D image data is an essential tool to increase the precision and safety of these operations. However, the plan’s effectiveness in the intra-operative workflow is challenged by unpredictable patient and device positioning and complex registration protocols. Here, we develop and analyze a multi-stage algorithm that combines deep learning-based anatomical feature detection and geometric post-processing to enable accurate pre- and intra-operative surgery planning on 2D X-ray images. The algorithm allows granular control over each element of the planning geometry, enabling real-time adjustments directly in the operating room (OR). In the method evaluation of three ligament reconstruction tasks effect on the knee joint, we found high spatial precision in drilling point localization (ε<2.9mm) and low angulation errors for k-wire instrumentation (ε<0.75) on 38 diagnostic radiographs. Comparable precision was demonstrated in 15 complex intra-operative trauma cases suffering from strong implant overlap and multi-anatomy exposure. Furthermore, we found that the diverse feature detection tasks can be efficiently solved with a multi-task network topology, improving precision over the single-task case. Our platform will help overcome the limitations of current clinical practice and foster surgical plan generation and adjustment directly in the OR, ultimately motivating the development of novel 2D planning guidelines. Full article
Show Figures

Graphical abstract

13 pages, 4255 KiB  
Article
Example-Based Multispectral Photometric Stereo for Multi-Colored Surfaces
by Daisuke Miyazaki and Kazuya Uegomori
J. Imaging 2022, 8(4), 107; https://doi.org/10.3390/jimaging8040107 - 11 Apr 2022
Cited by 1 | Viewed by 2527
Abstract
A photometric stereo needs three images taken under three different light directions lit one by one, while a color photometric stereo needs only one image taken under three different lights lit at the same time with different light directions and different colors. As [...] Read more.
A photometric stereo needs three images taken under three different light directions lit one by one, while a color photometric stereo needs only one image taken under three different lights lit at the same time with different light directions and different colors. As a result, a color photometric stereo can obtain the surface normal of a dynamically moving object from a single image. However, the conventional color photometric stereo cannot estimate a multicolored object due to the colored illumination. This paper uses an example-based photometric stereo to solve the problem of the color photometric stereo. The example-based photometric stereo searches the surface normal from the database of the images of known shapes. Color photometric stereos suffer from mathematical difficulty, and they add many assumptions and constraints; however, the example-based photometric stereo is free from such mathematical problems. The process of our method is pixelwise; thus, the estimated surface normal is not oversmoothed, unlike existing methods that use smoothness constraints. To demonstrate the effectiveness of this study, a measurement device that can realize the multispectral photometric stereo method with sixteen colors is employed instead of the classic color photometric stereo method with three colors. Full article
Show Figures

Figure 1

21 pages, 2195 KiB  
Article
Adaptive Real-Time Object Detection for Autonomous Driving Systems
by Maryam Hemmati, Morteza Biglari-Abhari and Smail Niar
J. Imaging 2022, 8(4), 106; https://doi.org/10.3390/jimaging8040106 - 11 Apr 2022
Cited by 3 | Viewed by 2426
Abstract
Accurate and reliable detection is one of the main tasks of Autonomous Driving Systems (ADS). While detecting the obstacles on the road during various environmental circumstances add to the reliability of ADS, it results in more intensive computations and more complicated systems. The [...] Read more.
Accurate and reliable detection is one of the main tasks of Autonomous Driving Systems (ADS). While detecting the obstacles on the road during various environmental circumstances add to the reliability of ADS, it results in more intensive computations and more complicated systems. The stringent real-time requirements of ADS, resource constraints, and energy efficiency considerations add to the design complications. This work presents an adaptive system that detects pedestrians and vehicles in different lighting conditions on the road. We take a hardware-software co-design approach on Zynq UltraScale+ MPSoC and develop a dynamically reconfigurable ADS that employs hardware accelerators for pedestrian and vehicle detection and adapts its detection method to the environment lighting conditions. The results show that the system maintains real-time performance and achieves adaptability with minimal resource overhead. Full article
(This article belongs to the Special Issue Image Processing Using FPGAs 2021)
Show Figures

Figure 1

20 pages, 1154 KiB  
Article
Face Attribute Estimation Using Multi-Task Convolutional Neural Network
by Hiroya Kawai, Koichi Ito and Takafumi Aoki
J. Imaging 2022, 8(4), 105; https://doi.org/10.3390/jimaging8040105 - 10 Apr 2022
Cited by 1 | Viewed by 3048
Abstract
Face attribute estimation can be used for improving the accuracy of face recognition, customer analysis in marketing, image retrieval, video surveillance, and criminal investigation. The major methods for face attribute estimation are based on Convolutional Neural Networks (CNNs) that solve face attribute estimation [...] Read more.
Face attribute estimation can be used for improving the accuracy of face recognition, customer analysis in marketing, image retrieval, video surveillance, and criminal investigation. The major methods for face attribute estimation are based on Convolutional Neural Networks (CNNs) that solve face attribute estimation as a multiple two-class classification problem. Although one feature extractor should be used for each attribute to explore the accuracy of attribute estimation, in most cases, one feature extractor is shared to estimate all face attributes for the parameter efficiency. This paper proposes a face attribute estimation method using Merged Multi-CNN (MM-CNN) to automatically optimize CNN structures for solving multiple binary classification problems to improve parameter efficiency and accuracy in face attribute estimation. We also propose a parameter reduction method called Convolutionalization for Parameter Reduction (CPR), which removes all fully connected layers from MM-CNNs. Through a set of experiments using the CelebA and LFW-a datasets, we demonstrate that MM-CNN with CPR exhibits higher efficiency of face attribute estimation in terms of estimation accuracy and the number of weight parameters than conventional methods. Full article
(This article belongs to the Special Issue Intelligent Media Processing)
Show Figures

Figure 1

24 pages, 10116 KiB  
Article
Explainable Multimedia Feature Fusion for Medical Applications
by Stefan Wagenpfeil, Paul Mc Kevitt, Abbas Cheddad and Matthias Hemmje
J. Imaging 2022, 8(4), 104; https://doi.org/10.3390/jimaging8040104 - 8 Apr 2022
Cited by 3 | Viewed by 2563
Abstract
Due to the exponential growth of medical information in the form of, e.g., text, images, Electrocardiograms (ECGs), X-rays, and multimedia, the management of a patient’s data has become a huge challenge. In particular, the extraction of features from various different formats and their [...] Read more.
Due to the exponential growth of medical information in the form of, e.g., text, images, Electrocardiograms (ECGs), X-rays, and multimedia, the management of a patient’s data has become a huge challenge. In particular, the extraction of features from various different formats and their representation in a homogeneous way are areas of interest in medical applications. Multimedia Information Retrieval (MMIR) frameworks, like the Generic Multimedia Analysis Framework (GMAF), can contribute to solving this problem, when adapted to special requirements and modalities of medical applications. In this paper, we demonstrate how typical multimedia processing techniques can be extended and adapted to medical applications and how these applications benefit from employing a Multimedia Feature Graph (MMFG) and specialized, efficient indexing structures in the form of Graph Codes. These Graph Codes are transformed to feature relevant Graph Codes by employing a modified Term Frequency Inverse Document Frequency (TFIDF) algorithm, which further supports value ranges and Boolean operations required in the medical context. On this basis, various metrics for the calculation of similarity, recommendations, and automated inferencing and reasoning can be applied supporting the field of diagnostics. Finally, the presentation of these new facilities in the form of explainability is introduced and demonstrated. Thus, in this paper, we show how Graph Codes contribute new querying options for diagnosis and how Explainable Graph Codes can help to readily understand medical multimedia formats. Full article
(This article belongs to the Special Issue Intelligent Strategies for Medical Image Analysis)
Show Figures

Figure 1

17 pages, 13304 KiB  
Article
A Hybrid Method for 3D Reconstruction of MR Images
by Loubna Lechelek, Sebastien Horna, Rita Zrour, Mathieu Naudin and Carole Guillevin
J. Imaging 2022, 8(4), 103; https://doi.org/10.3390/jimaging8040103 - 7 Apr 2022
Cited by 5 | Viewed by 2908
Abstract
Three-dimensional surface reconstruction is a well-known task in medical imaging. In procedures for intervention or radiation treatment planning, the generated models should be accurate and reflect the natural appearance. Traditional methods for this task, such as Marching Cubes, use smoothing post processing to [...] Read more.
Three-dimensional surface reconstruction is a well-known task in medical imaging. In procedures for intervention or radiation treatment planning, the generated models should be accurate and reflect the natural appearance. Traditional methods for this task, such as Marching Cubes, use smoothing post processing to reduce staircase artifacts from mesh generation and exhibit the natural look. However, smoothing algorithms often reduce the quality and degrade the accuracy. Other methods, such as MPU implicits, based on adaptive implicit functions, inherently produce smooth 3D models. However, the integration in the implicit functions of both smoothness and accuracy of the shape approximation may impact the precision of the reconstruction. Having these limitations in mind, we propose a hybrid method for 3D reconstruction of MR images. This method is based on a parallel Marching Cubes algorithm called Flying Edges (FE) and Multi-level Partition of Unity (MPU) implicits. We aim to combine the robustness of the Marching Cubes algorithm with the smooth implicit curve tracking enabled by the use of implicit models in order to provide higher geometry precision. Towards this end, the regions that closely fit to the segmentation data, and thus regions that are not impacted by reconstruction issues, are first extracted from both methods. These regions are then merged and used to reconstruct the final model. Experimental studies were performed on a number of MRI datasets, providing images and error statistics generated from our results. The results obtained show that our method reduces the geometric errors of the reconstructed surfaces when compared to the MPU and FE approaches, producing a more accurate 3D reconstruction. Full article
(This article belongs to the Special Issue Geometry Reconstruction from Images)
Show Figures

Figure 1

18 pages, 6116 KiB  
Article
Novel Hypertrophic Cardiomyopathy Diagnosis Index Using Deep Features and Local Directional Pattern Techniques
by Anjan Gudigar, U. Raghavendra, Jyothi Samanth, Chinmay Dharmik, Mokshagna Rohit Gangavarapu, Krishnananda Nayak, Edward J. Ciaccio, Ru-San Tan, Filippo Molinari and U. Rajendra Acharya
J. Imaging 2022, 8(4), 102; https://doi.org/10.3390/jimaging8040102 - 6 Apr 2022
Cited by 6 | Viewed by 3115
Abstract
Hypertrophic cardiomyopathy (HCM) is a genetic disorder that exhibits a wide spectrum of clinical presentations, including sudden death. Early diagnosis and intervention may avert the latter. Left ventricular hypertrophy on heart imaging is an important diagnostic criterion for HCM, and the most common [...] Read more.
Hypertrophic cardiomyopathy (HCM) is a genetic disorder that exhibits a wide spectrum of clinical presentations, including sudden death. Early diagnosis and intervention may avert the latter. Left ventricular hypertrophy on heart imaging is an important diagnostic criterion for HCM, and the most common imaging modality is heart ultrasound (US). The US is operator-dependent, and its interpretation is subject to human error and variability. We proposed an automated computer-aided diagnostic tool to discriminate HCM from healthy subjects on US images. We used a local directional pattern and the ResNet-50 pretrained network to classify heart US images acquired from 62 known HCM patients and 101 healthy subjects. Deep features were ranked using Student’s t-test, and the most significant feature (SigFea) was identified. An integrated index derived from the simulation was defined as 100·log10(SigFea/2)  in each subject, and a diagnostic threshold value was empirically calculated as the mean of the minimum and maximum integrated indices among HCM and healthy subjects, respectively. An integrated index above a threshold of 0.5 separated HCM from healthy subjects with 100% accuracy in our test dataset. Full article
Show Figures

Figure 1

19 pages, 10222 KiB  
Article
Machine-Learning-Based Real-Time Multi-Camera Vehicle Tracking and Travel-Time Estimation
by Xiaohui Huang, Pan He, Anand Rangarajan and Sanjay Ranka
J. Imaging 2022, 8(4), 101; https://doi.org/10.3390/jimaging8040101 - 6 Apr 2022
Cited by 3 | Viewed by 3060
Abstract
Travel-time estimation of traffic flow is an important problem with critical implications for traffic congestion analysis. We developed techniques for using intersection videos to identify vehicle trajectories across multiple cameras and analyze corridor travel time. Our approach consists of (1) multi-object single-camera tracking, [...] Read more.
Travel-time estimation of traffic flow is an important problem with critical implications for traffic congestion analysis. We developed techniques for using intersection videos to identify vehicle trajectories across multiple cameras and analyze corridor travel time. Our approach consists of (1) multi-object single-camera tracking, (2) vehicle re-identification among different cameras, (3) multi-object multi-camera tracking, and (4) travel-time estimation. We evaluated the proposed framework on real intersections in Florida with pan and fisheye cameras. The experimental results demonstrate the viability and effectiveness of our method. Full article
(This article belongs to the Special Issue Multi-Object Tracking)
Show Figures

Figure 1

10 pages, 1030 KiB  
Review
Cardiac Magnetic Resonance Imaging in Immune Check-Point Inhibitor Myocarditis: A Systematic Review
by Luca Arcari, Giacomo Tini, Giovanni Camastra, Federica Ciolina, Domenico De Santis, Domitilla Russo, Damiano Caruso, Massimiliano Danti and Luca Cacciotti
J. Imaging 2022, 8(4), 99; https://doi.org/10.3390/jimaging8040099 - 5 Apr 2022
Cited by 5 | Viewed by 2729
Abstract
Immune checkpoint inhibitors (ICIs) are a family of anticancer drugs in which the immune response elicited against the tumor may involve other organs, including the heart. Cardiac magnetic resonance (CMR) imaging is increasingly used in the diagnostic work-up of myocardial inflammation; recently, several [...] Read more.
Immune checkpoint inhibitors (ICIs) are a family of anticancer drugs in which the immune response elicited against the tumor may involve other organs, including the heart. Cardiac magnetic resonance (CMR) imaging is increasingly used in the diagnostic work-up of myocardial inflammation; recently, several studies investigated the use of CMR in patients with ICI-myocarditis (ICI-M). The aim of the present systematic review is to summarize the available evidence on CMR findings in ICI-M. We searched electronic databases for relevant publications; after screening, six studies were selected, including 166 patients from five cohorts, and further 86 patients from a sub-analysis that were targeted for a tissue mapping assessment. CMR revealed mostly preserved left ventricular ejection fraction; edema prevalence ranged from 9% to 60%; late gadolinium enhancement (LGE) prevalence ranged from 23% to 83%. T1 and T2 mapping assessment were performed in 108 and 104 patients, respectively. When available, the comparison of CMR with endomyocardial biopsy revealed partial agreement between techniques and was higher for native T1 mapping amongst imaging biomarkers. The prognostic assessment was inconsistently assessed; CMR variables independently associated with the outcome included decreasing LVEF and increasing native T1. In conclusion, CMR findings in ICI-M include myocardial dysfunction, edema and fibrosis, though less evident than in more classic forms of myocarditis; native T1 mapping retained the higher concordance with EMB and significant prognostic value. Full article
(This article belongs to the Special Issue Intelligent Strategies for Medical Image Analysis)
Show Figures

Figure 1

11 pages, 1977 KiB  
Article
Transverse Analysis of Maxilla and Mandible in Adults with Normal Occlusion: A Cone Beam Computed Tomography Study
by Kyung Jin Lee, Hyeran Helen Jeon, Normand Boucher and Chun-Hsi Chung
J. Imaging 2022, 8(4), 100; https://doi.org/10.3390/jimaging8040100 - 5 Apr 2022
Cited by 4 | Viewed by 3120
Abstract
Objectives: To study the transverse widths of maxilla and mandible and their relationship with the inclination of first molars. Materials and Methods: Fifty-six untreated adults (12 males, 44 females) with normal occlusion were included. On each Cone Beam Computed Tomography (CBCT) image of [...] Read more.
Objectives: To study the transverse widths of maxilla and mandible and their relationship with the inclination of first molars. Materials and Methods: Fifty-six untreated adults (12 males, 44 females) with normal occlusion were included. On each Cone Beam Computed Tomography (CBCT) image of the subject, inter-buccal and inter-lingual bone widths were measured at the levels of hard palate, alveolar crest and furcation of the first molars, and maxillomandibular width differentials were calculated. In addition, the buccolingual inclination of each first molar was measured and its correlation with the maxillomandibular width differential was tested. Results: At the furcation level of the first molar, the maxillary inter-buccal bone width was more than the mandibular inter-buccal bone width by 1.1 ± 4.5 mm for males and 1.6 ± 2.9 mm for females; the mandibular inter-lingual bone width was more than the maxillary inter-lingual bone width by 1.3 ± 3.6 mm for males and 0.3 ± 3.2 mm for females. For females, there was a negative correlation between the maxillomandibular inter-lingual bone differential and maxillary first molar buccal inclination (p < 0.05), and a positive correlation between the maxillomandibular inter-lingual bone differential and mandibular first molar lingual inclination (p < 0.05). Conclusions: This is a randomized clinical study on transverse analysis of maxilla and mandible in adults with normal occlusion using CBCTs. On average: (1) At the furcation level of the first molars, the maxillary inter-buccal bone width was slightly wider than mandibular inter-buccal bone width; whereas the mandibular inter-lingual bone width was slightly wider than maxillary inter-lingual bone width; (2) A statistically significant correlation existed between the maxillomandibular transverse skeletal differentials and molar inclinations. Full article
(This article belongs to the Special Issue New Frontiers of Advanced Imaging in Dentistry)
Show Figures

Figure 1

48 pages, 5076 KiB  
Review
A Comparative Review on Applications of Different Sensors for Sign Language Recognition
by Muhammad Saad Amin, Syed Tahir Hussain Rizvi and Md. Murad Hossain
J. Imaging 2022, 8(4), 98; https://doi.org/10.3390/jimaging8040098 - 2 Apr 2022
Cited by 21 | Viewed by 6918
Abstract
Sign language recognition is challenging due to the lack of communication between normal and affected people. Many social and physiological impacts are created due to speaking or hearing disability. A lot of different dimensional techniques have been proposed previously to overcome this gap. [...] Read more.
Sign language recognition is challenging due to the lack of communication between normal and affected people. Many social and physiological impacts are created due to speaking or hearing disability. A lot of different dimensional techniques have been proposed previously to overcome this gap. A sensor-based smart glove for sign language recognition (SLR) proved helpful to generate data based on various hand movements related to specific signs. A detailed comparative review of all types of available techniques and sensors used for sign language recognition was presented in this article. The focus of this paper was to explore emerging trends and strategies for sign language recognition and to point out deficiencies in existing systems. This paper will act as a guide for other researchers to understand all materials and techniques like flex resistive sensor-based, vision sensor-based, or hybrid system-based technologies used for sign language until now. Full article
Show Figures

Figure 1

16 pages, 2855 KiB  
Article
Machine Learning for Early Parkinson’s Disease Identification within SWEDD Group Using Clinical and DaTSCAN SPECT Imaging Features
by Hajer Khachnaoui, Nawres Khlifa and Rostom Mabrouk
J. Imaging 2022, 8(4), 97; https://doi.org/10.3390/jimaging8040097 - 2 Apr 2022
Cited by 17 | Viewed by 3164
Abstract
Early Parkinson’s Disease (PD) diagnosis is a critical challenge in the treatment process. Meeting this challenge allows appropriate planning for patients. However, Scan Without Evidence of Dopaminergic Deficit (SWEDD) is a heterogeneous group of PD patients and Healthy Controls (HC) in clinical and [...] Read more.
Early Parkinson’s Disease (PD) diagnosis is a critical challenge in the treatment process. Meeting this challenge allows appropriate planning for patients. However, Scan Without Evidence of Dopaminergic Deficit (SWEDD) is a heterogeneous group of PD patients and Healthy Controls (HC) in clinical and imaging features. The application of diagnostic tools based on Machine Learning (ML) comes into play here as they are capable of distinguishing between HC subjects and PD patients within an SWEDD group. In the present study, three ML algorithms were used to separate PD patients from HC within an SWEDD group. Data of 548 subjects were firstly analyzed by Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) techniques. Using the best reduction technique result, we built the following clustering models: Density-Based Spatial (DBSCAN), K-means and Hierarchical Clustering. According to our findings, LDA performs better than PCA; therefore, LDA was used as input for the clustering models. The different models’ performances were assessed by comparing the clustering algorithms outcomes with the ground truth after a follow-up. Hierarchical Clustering surpassed DBSCAN and K-means algorithms by 64%, 78.13% and 38.89% in terms of accuracy, sensitivity and specificity. The proposed method demonstrated the suitability of ML models to distinguish PD patients from HC subjects within an SWEDD group. Full article
Show Figures

Figure 1

16 pages, 1428 KiB  
Article
Convolutional Neural Networks for the Identification of African Lions from Individual Vocalizations
by Martino Trapanotto, Loris Nanni, Sheryl Brahnam and Xiang Guo
J. Imaging 2022, 8(4), 96; https://doi.org/10.3390/jimaging8040096 - 1 Apr 2022
Cited by 6 | Viewed by 3243
Abstract
The classification of vocal individuality for passive acoustic monitoring (PAM) and census of animals is becoming an increasingly popular area of research. Nearly all studies in this field of inquiry have relied on classic audio representations and classifiers, such as Support Vector Machines [...] Read more.
The classification of vocal individuality for passive acoustic monitoring (PAM) and census of animals is becoming an increasingly popular area of research. Nearly all studies in this field of inquiry have relied on classic audio representations and classifiers, such as Support Vector Machines (SVMs) trained on spectrograms or Mel-Frequency Cepstral Coefficients (MFCCs). In contrast, most current bioacoustic species classification exploits the power of deep learners and more cutting-edge audio representations. A significant reason for avoiding deep learning in vocal identity classification is the tiny sample size in the collections of labeled individual vocalizations. As is well known, deep learners require large datasets to avoid overfitting. One way to handle small datasets with deep learning methods is to use transfer learning. In this work, we evaluate the performance of three pretrained CNNs (VGG16, ResNet50, and AlexNet) on a small, publicly available lion roar dataset containing approximately 150 samples taken from five male lions. Each of these networks is retrained on eight representations of the samples: MFCCs, spectrogram, and Mel spectrogram, along with several new ones, such as VGGish and stockwell, and those based on the recently proposed LM spectrogram. The performance of these networks, both individually and in ensembles, is analyzed and corroborated using the Equal Error Rate and shown to surpass previous classification attempts on this dataset; the best single network achieved over 95% accuracy and the best ensembles over 98% accuracy. The contributions this study makes to the field of individual vocal classification include demonstrating that it is valuable and possible, with caution, to use transfer learning with single pretrained CNNs on the small datasets available for this problem domain. We also make a contribution to bioacoustics generally by offering a comparison of the performance of many state-of-the-art audio representations, including for the first time the LM spectrogram and stockwell representations. All source code for this study is available on GitHub. Full article
(This article belongs to the Special Issue Computer Vision and Deep Learning: Trends and Applications)
Show Figures

Figure 1

8 pages, 1013 KiB  
Article
Spiky: An ImageJ Plugin for Data Analysis of Functional Cardiac and Cardiomyocyte Studies
by Côme Pasqualin, François Gannier, Angèle Yu, David Benoist, Ian Findlay, Romain Bordy, Pierre Bredeloux and Véronique Maupoil
J. Imaging 2022, 8(4), 95; https://doi.org/10.3390/jimaging8040095 - 1 Apr 2022
Cited by 6 | Viewed by 3659
Abstract
Introduction and objective: Nowadays, investigations of heart physiology and pathophysiology rely more and more upon image analysis, whether for the detection and characterization of events in single cells or for the mapping of events and their characteristics across an entire tissue. These investigations [...] Read more.
Introduction and objective: Nowadays, investigations of heart physiology and pathophysiology rely more and more upon image analysis, whether for the detection and characterization of events in single cells or for the mapping of events and their characteristics across an entire tissue. These investigations require extensive skills in image analysis and/or expensive software, and their reproducibility may be a concern. Our objective was to build a robust, reliable and open-source software tool to quantify excitation–contraction related experimental data at multiple scales, from single isolated cells to the whole heart. Methods and results: A free and open-source ImageJ plugin, Spiky, was developed to detect and analyze peaks in experimental data streams. It allows rapid and easy analysis of action potentials, intracellular calcium transient and contraction data from cardiac research experiments. As shown in the provided examples, both classical bi-dimensional data (XT signals) and video data obtained from confocal microscopy and optical mapping experiments (XYT signals) can be analyzed. Spiky was written in ImageJ Macro Language and JAVA, and works under Windows, Mac and Linux operating systems. Conclusion: Spiky provides a complete working interface to process and analyze cardiac physiology research data. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

16 pages, 21314 KiB  
Article
Imaging PPG for In Vivo Human Tissue Perfusion Assessment during Surgery
by Marco Lai, Stefan D. van der Stel, Harald C. Groen, Mark van Gastel, Koert F. D. Kuhlmann, Theo J. M. Ruers and Benno H. W. Hendriks
J. Imaging 2022, 8(4), 94; https://doi.org/10.3390/jimaging8040094 - 31 Mar 2022
Cited by 10 | Viewed by 2941
Abstract
Surgical excision is the golden standard for treatment of intestinal tumors. In this surgical procedure, inadequate perfusion of the anastomosis can lead to postoperative complications, such as anastomotic leakages. Imaging photoplethysmography (iPPG) can potentially provide objective and real-time feedback of the perfusion status [...] Read more.
Surgical excision is the golden standard for treatment of intestinal tumors. In this surgical procedure, inadequate perfusion of the anastomosis can lead to postoperative complications, such as anastomotic leakages. Imaging photoplethysmography (iPPG) can potentially provide objective and real-time feedback of the perfusion status of tissues. This feasibility study aims to evaluate an iPPG acquisition system during intestinal surgeries to detect the perfusion levels of the microvasculature tissue bed in different perfusion conditions. This feasibility study assesses three patients that underwent resection of a portion of the small intestine. Data was acquired from fully perfused, non-perfused and anastomosis parts of the intestine during different phases of the surgical procedure. Strategies for limiting motion and noise during acquisition were implemented. iPPG perfusion maps were successfully extracted from the intestine microvasculature, demonstrating that iPPG can be successfully used for detecting perturbations and perfusion changes in intestinal tissues during surgery. This study provides proof of concept for iPPG to detect changes in organ perfusion levels. Full article
Show Figures

Figure 1

34 pages, 3852 KiB  
Article
Unified Probabilistic Deep Continual Learning through Generative Replay and Open Set Recognition
by Martin Mundt, Iuliia Pliushch, Sagnik Majumder, Yongwon Hong and Visvanathan Ramesh
J. Imaging 2022, 8(4), 93; https://doi.org/10.3390/jimaging8040093 - 31 Mar 2022
Cited by 15 | Viewed by 3837
Abstract
Modern deep neural networks are well known to be brittle in the face of unknown data instances and recognition of the latter remains a challenge. Although it is inevitable for continual-learning systems to encounter such unseen concepts, the corresponding literature appears to nonetheless [...] Read more.
Modern deep neural networks are well known to be brittle in the face of unknown data instances and recognition of the latter remains a challenge. Although it is inevitable for continual-learning systems to encounter such unseen concepts, the corresponding literature appears to nonetheless focus primarily on alleviating catastrophic interference with learned representations. In this work, we introduce a probabilistic approach that connects these perspectives based on variational inference in a single deep autoencoder model. Specifically, we propose to bound the approximate posterior by fitting regions of high density on the basis of correctly classified data points. These bounds are shown to serve a dual purpose: unseen unknown out-of-distribution data can be distinguished from already trained known tasks towards robust application. Simultaneously, to retain already acquired knowledge, a generative replay process can be narrowed to strictly in-distribution samples, in order to significantly alleviate catastrophic interference. Full article
(This article belongs to the Special Issue Continual Learning in Computer Vision: Theory and Applications)
Show Figures

Figure 1

15 pages, 29475 KiB  
Article
A New Preclinical Decision Support System Based on PET Radiomics: A Preliminary Study on the Evaluation of an Innovative 64Cu-Labeled Chelator in Mouse Models
by Viviana Benfante, Alessandro Stefano, Albert Comelli, Paolo Giaccone, Francesco Paolo Cammarata, Selene Richiusa, Fabrizio Scopelliti, Marco Pometti, Milene Ficarra, Sebastiano Cosentino, Marcello Lunardon, Francesca Mastrotto, Alberto Andrighetto, Antonino Tuttolomondo, Rosalba Parenti, Massimo Ippolito and Giorgio Russo
J. Imaging 2022, 8(4), 92; https://doi.org/10.3390/jimaging8040092 - 30 Mar 2022
Cited by 17 | Viewed by 3191
Abstract
The 64Cu-labeled chelator was analyzed in vivo by positron emission tomography (PET) imaging to evaluate its biodistribution in a murine model at different acquisition times. For this purpose, nine 6-week-old female Balb/C nude strain mice underwent micro-PET imaging at three different time [...] Read more.
The 64Cu-labeled chelator was analyzed in vivo by positron emission tomography (PET) imaging to evaluate its biodistribution in a murine model at different acquisition times. For this purpose, nine 6-week-old female Balb/C nude strain mice underwent micro-PET imaging at three different time points after 64Cu-labeled chelator injection. Specifically, the mice were divided into group 1 (acquisition 1 h after [64Cu] chelator administration, n = 3 mice), group 2 (acquisition 4 h after [64Cu]chelator administration, n = 3 mice), and group 3 (acquisition 24 h after [64Cu] chelator administration, n = 3 mice). Successively, all PET studies were segmented by means of registration with a standard template space (3D whole-body Digimouse atlas), and 108 radiomics features were extracted from seven organs (namely, heart, bladder, stomach, liver, spleen, kidney, and lung) to investigate possible changes over time in [64Cu]chelator biodistribution. The one-way analysis of variance and post hoc Tukey Honestly Significant Difference test revealed that, while heart, stomach, spleen, kidney, and lung districts showed a very low percentage of radiomics features with significant variations (p-value < 0.05) among the three groups of mice, a large number of features (greater than 60% and 50%, respectively) that varied significantly between groups were observed in bladder and liver, indicating a different in vivo uptake of the 64Cu-labeled chelator over time. The proposed methodology may improve the method of calculating the [64Cu]chelator biodistribution and open the way towards a decision support system in the field of new radiopharmaceuticals used in preclinical imaging trials. Full article
(This article belongs to the Special Issue Radiomics and Texture Analysis in Medical Imaging)
Show Figures

Figure 1

20 pages, 918 KiB  
Systematic Review
Augmented Reality Games and Presence: A Systematic Review
by Anabela Marto and Alexandrino Gonçalves
J. Imaging 2022, 8(4), 91; https://doi.org/10.3390/jimaging8040091 - 29 Mar 2022
Cited by 14 | Viewed by 4957
Abstract
The sense of presence in augmented reality (AR) has been studied by multiple researchers through diverse applications and strategies. In addition to the valuable information provided to the scientific community, new questions keep being raised. These approaches vary from following the standards from [...] Read more.
The sense of presence in augmented reality (AR) has been studied by multiple researchers through diverse applications and strategies. In addition to the valuable information provided to the scientific community, new questions keep being raised. These approaches vary from following the standards from virtual reality to ascertaining the presence of users’ experiences and new proposals for evaluating presence that specifically target AR environments. It is undeniable that the idea of evaluating presence across AR may be overwhelming due to the different scenarios that may be possible, whether this regards technological devices—from immersive AR headsets to the small screens of smartphones—or the amount of virtual information that is being added to the real scenario. Taking into account the recent literature that has addressed the sense of presence in AR as a true challenge given the diversity of ways that AR can be experienced, this study proposes a specific scope to address presence and other related forms of dimensions such as immersion, engagement, embodiment, or telepresence, when AR is used in games. This systematic review was conducted following the PRISMA methodology, carefully analysing all studies that reported visual games that include AR activities and somehow included presence data—or related dimensions that may be referred to as immersion-related feelings, analysis or results. This study clarifies what dimensions of presence are being considered and evaluated in AR games, how presence-related variables have been evaluated, and what the major research findings are. For a better understanding of these approaches, this study takes note of what devices are being used for the AR experience when immersion-related feelings are one of the behaviours that are considered in their evaluations, and discusses to what extent these feelings in AR games affect the player’s other behaviours. Full article
(This article belongs to the Special Issue Advanced Scene Perception for Augmented Reality)
Show Figures

Figure 1

12 pages, 5773 KiB  
Article
Object Categorization Capability of Psychological Potential Field in Perceptual Assessment Using Line-Drawing Images
by Naoyuki Awano and Yuki Hayashi
J. Imaging 2022, 8(4), 90; https://doi.org/10.3390/jimaging8040090 - 26 Mar 2022
Cited by 1 | Viewed by 2006
Abstract
Affective/cognitive engineering investigations typically require the quantitative assessment of object perception. Recent research has suggested that certain perceptions of object categorization can be derived from human eye fixation and that color images and line drawings induce similar neural activities. Line drawings contain less [...] Read more.
Affective/cognitive engineering investigations typically require the quantitative assessment of object perception. Recent research has suggested that certain perceptions of object categorization can be derived from human eye fixation and that color images and line drawings induce similar neural activities. Line drawings contain less information than color images; therefore, line drawings are expected to simplify the investigations of object perception. The psychological potential field (PPF), which is a psychological feature, is an image feature of line drawings. On the basis of the PPF, the possibility that the general human perception of object categorization can be assessed from the similarity to fixation maps (FMs) generated from human eye fixations has been reported. However, this may be due to chance because image features other than the PPF have not been compared with FMs. This study examines the potential and effectiveness of the PPF by comparing its performance with that of other image features in terms of the similarity to FMs. The results show that the PPF shows the ideal performance for assessing the perception of object categorization. In particular, the PPF effectively distinguishes between animal and nonanimal targets; however, real-time assessment is difficult. Full article
(This article belongs to the Special Issue Human Attention and Visual Cognition)
Show Figures

Figure 1

21 pages, 399 KiB  
Article
Union-Retire for Connected Components Analysis on FPGA
by Donald G. Bailey and Michael J. Klaiber
J. Imaging 2022, 8(4), 89; https://doi.org/10.3390/jimaging8040089 - 24 Mar 2022
Cited by 1 | Viewed by 1781
Abstract
The Union-Retire CCA (UR-CCA) algorithm started a new paradigm for connected components analysis. Instead of using directed tree structures, UR-CCA focuses on connectivity. This algorithmic change leads to a reduction in required memory, with no end-of-row processing overhead. In this paper we describe [...] Read more.
The Union-Retire CCA (UR-CCA) algorithm started a new paradigm for connected components analysis. Instead of using directed tree structures, UR-CCA focuses on connectivity. This algorithmic change leads to a reduction in required memory, with no end-of-row processing overhead. In this paper we describe a hardware architecture based on UR-CCA and its realisation on an FPGA. The memory bandwidth and pipelining challenges of hardware UR-CCA are analysed and resolved. It is shown that up to 36% of memory resources can be saved using the proposed architecture. This translates directly to a smaller device for an FPGA implementation. Full article
(This article belongs to the Special Issue Image Processing Using FPGAs 2021)
Show Figures

Figure 1

14 pages, 8913 KiB  
Article
YOLOv4-Based CNN Model versus Nested Contours Algorithm in the Suspicious Lesion Detection on the Mammography Image: A Direct Comparison in the Real Clinical Settings
by Alexey Kolchev, Dmitry Pasynkov, Ivan Egoshin, Ivan Kliouchkin, Olga Pasynkova and Dmitrii Tumakov
J. Imaging 2022, 8(4), 88; https://doi.org/10.3390/jimaging8040088 - 24 Mar 2022
Cited by 11 | Viewed by 3446
Abstract
Background: We directly compared the mammography image processing results obtained with the help of the YOLOv4 convolutional neural network (CNN) model versus those obtained with the help of the NCA-based nested contours algorithm model. Method: We used 1080 images to train the YOLOv4, [...] Read more.
Background: We directly compared the mammography image processing results obtained with the help of the YOLOv4 convolutional neural network (CNN) model versus those obtained with the help of the NCA-based nested contours algorithm model. Method: We used 1080 images to train the YOLOv4, plus 100 images with proven breast cancer (BC) and 100 images with proven absence of BC to test both models. Results: the rates of true-positive, false-positive and false-negative outcomes were 60, 10 and 40, respectively, for YOLOv4, and 93, 63 and 7, respectively, for NCA. The sensitivities for the YOLOv4 and the NCA were comparable to each other for star-like lesions, masses with unclear borders, round- or oval-shaped masses with clear borders and partly visualized masses. On the contrary, the NCA was superior to the YOLOv4 in the case of asymmetric density and of changes invisible on the dense parenchyma background. Radiologists changed their earlier decisions in six cases per 100 for NCA. YOLOv4 outputs did not influence the radiologists’ decisions. Conclusions: in our set, NCA clinically significantly surpasses YOLOv4. Full article
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop