Next Article in Journal
Threshold-Responsive Colorimetric Sensing System for the Continuous Monitoring of Gases
Next Article in Special Issue
On-Device Execution of Deep Learning Models on HoloLens2 for Real-Time Augmented Reality Medical Applications
Previous Article in Journal
Editorial for Special Issue: Machine Health Monitoring and Fault Diagnosis Techniques
Previous Article in Special Issue
Magic Leap 1 versus Microsoft HoloLens 2 for the Visualization of 3D Content Obtained from Radiological Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Head-Mounted Projector for Manual Precision Tasks: Performance Assessment

1
EndoCAS Center for Computer-Assisted Surgery, University of Pisa, 56124 Pisa, Italy
2
Azienda Ospedaliero Universitaria Pisana, 56126 Pisa, Italy
3
Information Engineering Department, University of Pisa, 56126 Pisa, Italy
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3494; https://doi.org/10.3390/s23073494
Submission received: 13 February 2023 / Revised: 22 March 2023 / Accepted: 24 March 2023 / Published: 27 March 2023

Abstract

:
The growing interest in augmented reality applications has led to an in-depth look at the performance of head-mounted displays and their testing in numerous domains. Other devices for augmenting the real world with virtual information are presented less frequently and usually focus on the description of the device rather than on its performance analysis. This is the case of projected augmented reality, which, compared to head-worn AR displays, offers the advantages of being simultaneously accessible by multiple users whilst preserving user awareness of the environment and feeling of immersion. This work provides a general evaluation of a custom-made head-mounted projector for the aid of precision manual tasks through an experimental protocol designed for investigating spatial and temporal registration and their combination. The results of the tests show that the accuracy ( 0.6 ± 0.1 mm of spatial registration error) and motion-to-photon latency ( 113 ± 12 ms) make the proposed solution suitable for guiding precision tasks.

1. Introduction

In recent years, the potential of Augmented Reality (AR) technology to enhance the real world with auxiliary virtual information has become increasingly prominent. To this end, Head Mounted Displays (HMDs) are by far the most studied devices. Their application has been proposed in various fields ranging from industrial quality control and assembly [1,2], education [3], military [4,5], and surgery [6,7] to entertainment [8]. In the literature, the characteristics of HMDs have been extensively studied and a few drawbacks have emerged such as the difficulty in achieving a feeling of immersion, the loss of spatial awareness [9], and the development of motion sickness [10]. These drawbacks contribute to an increased cognitive load resulting from well-known technological pitfalls typical of commercial and/or laboratory prototypes of HMDs. Among the technological drawbacks that limit the success of AR HMDs in the consumer market, there is the presence of a narrow display field of view, the discrete and low display resolution, the low frame rate, the perceivable motion-to-photon latency, and the perceptual issues such as the vergence-accommodation conflict and the real-to-virtual focus rivalry. These limitations strongly impact the perceptual experience and result in maladaptations, such as cybersickness [11], other than producing safety concerns. In addition, these limits greatly hinder the widespread adoption of AR HMDs to support the accurate and efficient execution of highly challenging manual activities that must be carried out in the peripersonal space (i.e., for surgical guidance) [12], with alternative AR solutions that have been proposed to overcome them.
Projected AR, also called spatial AR, lends itself to preserving the user’s awareness of the environment and the feeling of immersion since it does not require any optical interface interposed between the user’s eyes and the surrounding world. In projected AR, physical and digital information are mixed directly into the real world via visible light projectors. As a result, a single projected AR device can be simultaneously experienced by multiple users. However, such devices are mostly static, as they are installed in a fixed position, hence they inevitably constrain the user’s freedom of movement. With the development of increasingly smaller and lighter projectors, new solutions have been proposed to overcome this problem. Nazarova et al., for instance, mounted a tracking and projection module on a collaborative robot to provide immersive and intuitive robot control to the user [13]. Hartmann’s team proposed an actuated Head-Mounted Projector (HMP) to increase the user’s field of view and foster collaboration and communication by providing a shared experience with other people [14]. MoSART, instead, identifies itself as an alternative to classic HMDs through the development of an “all-in-one” headset that includes projection and tracking systems to provide AR for 3D Interaction with tangible objects [15]. The critical limitation of these systems is widely identified as the motion-to-photon (MTP) latency.
The term MTP latency was originally used to denote the average time between the instant when a motion occurs and the instant when the corresponding rendering event gets displayed on the display (i.e., in front of the user’s eyes for an HMD). MTP latency has been extensively studied for HMDs and literature results reveal their major impact on the system’s usability [10] since high latency can lead to substantial dynamic registration errors between the real world and that virtual content that heavily compromise the immersiveness of the experience, and it can also generate motion sickness and user discomfort [16]. A study by Jerald and Whitton showed that a 5 ms latency is completely imperceptible to a user in HMDs [17]. The Microsoft HoloLens comes very close to this value as they compensate for the slow optical tracking constrained by tracking the camera’s image acquisition time and processing time by exploiting tracking data from IMU sensors mounted on the headset coupled with motion-prediction 2D warping mechanisms. The latency of Meta 2 is about 60 ms [15], whereas that of HoloLens 2 is just 9 ms [18]. The stability of the virtual content, on the other hand, comes in terms of jitter. Jitter can be extremely disturbing for a human observer and can impair the feeling of immersion by breaking the illusion of a mixture between the virtual and real world. Jitter for Meta 2 and HoloLens is 1.7 mm and 0.2 mm at rest, respectively. A jitter of 1.7 mm has been identified as the upper limit for HMDs [19]. The same data are not available for HMPs proposed in the literature, although latency and jitter are recognized as critical aspects in the use of such devices. In [14], only the projector latency of 116 ms is given, with no mention of latency due to tracking or rendering that necessarily worsens the final value. MoSART system, on the other hand, claims a latency of 60 ms and jitter of 0.08 mm (MSE) but the same paper does not detail how these values were measured.
In this work, we propose an HMP built by assembling commercial devices, and we provide a detailed description of the assessment of its spatial and temporal registration together with their influence on the usability of the worn system [20]. The spatial registration accuracy depends on the geometry behind the AR implementation process, including calibration, tracking mechanism, and modeling of the target structures. Temporal registration comes into play when the user or the tracked object moves and it corresponds to synchronized motion between real and virtual structures over time. MTP latency is a measure of temporal registration. Specifically, we here present a protocol for:
  • Measure the temporal registration error of the HMP in terms of MTP latency.
  • Measure the static spatial registration error in terms of accuracy by evaluating the distance between the actual position of the virtual content and the desired position for perfect alignment in static conditions.
  • Measure how these two errors combine when a user wears the device and uses it to augment a tracked structure in real time.
All measurements were taken within the working area accessible to a user who intends to perform manual tasks while wearing the HMP. This area is between 350 mm and 600 mm from the user’s head, with an average value of 475 mm (i.e., the area within the user’s peripersonal space).

2. Materials and Methods

First, we present the hardware specifics of the HMP together with a brief explanation of the software framework that implements the tracking and the rendering. Then, the experimental protocol employed to evaluate the spatial and temporal registration accuracy and the combined effect of MTP latency and virtual-to-real static registration is described during a user test.

2.1. The Custom-Made Head-Mounted Projector

The custom-made HMP includes a Leopard Imaging OV580 stereo camera rig and an Anybeam laser mini-projector. The adopted camera configuration is 2560 × 720 @ 60 fps, which corresponds to an angular resolution of about 3.6 arcmin/px. The AnyBeam projector offers the advantage of no need for focus adjustment, as this is ensured by the MEMS laser technology. Projector resolution and brightness are, respectively, 720p @ 60 fps and 150 lumens. The stereo rig and the projector are calibrated to estimate both their intrinsic and extrinsic parameters [21,22]. The stability of the reciprocal pose of the components is ensured by a 3D-printed support modeled in Creo Parametric 3.0. The same support features an elastic band that allows the device to be worn on the head for an overall weight of about 300 g. The HMP is plugged into the computing unit where the software framework runs, consisting of a laptop with Intel Core i7-9750H CPU@ 2.60 GHz, 6 cores, 12 GB RAM, and GeForce RTX 2060 GPU by Nvidia.

2.2. Augmented Reality Software Framework

The software framework that controls the HMP was thoroughly described in [23]. Here, we recall its main features:
  • The software, originally conceived for the deployment of AR applications on video see-through (VST) and optical see-through (OST) display interfaces, is capable also of supporting the deployment of AR applications on different typologies of AR projectors. The software features a non-distributed architecture, which makes it also compatible with embedded computing units.
  • The software framework is based on Compute Unified Device Architecture (CUDA by Nvidia) to leverage parallel computing over the multi-core GPU card.
  • The software supports in situ visualization of 3D structures, thanks to the employment of the open-source library VTK for 3D computer graphics, modeling, and volume rendering.
  • The software framework is highly configurable in terms of rendering and tracking capabilities.
  • The software features a robust inside-out video-based tracking algorithm based on OpenCV API 3.4.1,
The software framework is able to support in situ visualization of medical imaging data; the key function of the software, under projector AR modality, is to process the images grabbed by the stereo pair of RGB cameras. The grabbed frames of the real scene are processed to perform marker-based optical tracking, which requires the identification of the 3D position of the markers both in the target reference frame and camera reference frame. The projected AR device needs a dedicated calibration between the world-facing tracking cameras and the projector. For this purpose, the projector is considered as an inverse camera according to the procedure exposed in [24]. Such calibration routine outputs the intrinsic projection parameters of the projector and the relative pose between the left camera and the projector itself.

2.3. Experimental Protocol

The experimental protocol relies on an external device for the evaluation of spatial and temporal registration accuracy. We selected a GoPro Hero 5 Black camera that features either a 4000 × 3000 photo resolution or a frame rate of up to 240 fps for videos at a 1280 × 720 resolution. As shown in Table 1, we adopted the highest resolution configuration for the static spatial registration tests and the highest frame rate for the dynamic temporal registration tests. For both configurations, we calibrated the GoPro Hero 5 and derived the average angular resolution (the amplitude of the visual angle covered by each pixel in arcmin) to enable valid comparison of the results. Table 1 also reports the working depth in order to obtain an estimate of the accuracy in physical distances (i.e., in mm).

2.3.1. Spatial Registration

We evaluated the AR spatial registration accuracy as the on-image distance between the location of a real point and its virtual counterpart (i.e., we measured the registration error in terms of target visualization error). For this purpose, we adopted a spatial registration evaluation tool consisting of three spherical markers fixed on a plane featuring two elevated points. The tool was modeled using Creo Parametric 3.0 and printed by rapid prototyping with a 600 dpi X-Y resolution, 1600 dpi Z resolution, and 0.1 mm precision. The two target points were dyed in red to be easily identified in the images as they are the reference points for AR alignment error assessment. The experimental set-up is shown in Figure 1a along with the spatial registration evaluation tool. The HMP and the GoPro Hero 5 were mounted on a holder that allows their distance from the working plane to be varied. In this way, we evaluated the registration error over the entire range of working distances between 350 mm and 600 mm (with a 25 mm step), a range of distances compatible with manual tasks. For each distance, we used GoPro setting 1 to capture five images of the evaluation tool. The acquired sets of images were automatically processed in Matlab™(version R2020.a) to estimate the 2D Target Visualization Error, that is the offset, expressed in pixels, between the virtual and real objects in the image plane (i.e., the centroids of the virtual landmarks and those of the corresponding real ones). The images were processed in hue-saturation-value color space and the circular Hough transform was used to detect the position of the virtual and real landmarks: first, virtual landmarks were detected in each series of AR images; then, their real counterparts were detected in the real image. 3D-printed landmarks were searched within a region of interest centered on the position of the corresponding virtual landmarks. We varied the orientation of the evaluation tool for each image captured at the same distance in order to include tracking error variability in the analysis: the target tool was placed in five different random orientations with respect to the tracking camera to account for the effect of different camera-to-target perspectives on the overall registration error. Overall, we collected 50 images, and for each image, we evaluated the distance of the two projected points from their real counterparts for a total of 100 evaluations of the error. Figure 1b shows one of the frames captured for the study, with the real points and their projected counterparts highlighted after manual selection.

2.3.2. Temporal Registration

For the evaluation of MTP latency, we chose to adopt the protocol proposed in [25]. Specifically, we fixed the HMP and the GoPro camera on the holder at a distance of 475 mm from the working plane, selected as the mean working distance for manual tasks. We placed the spatial registration evaluation tool used in previous tests on the working plane. We measured the latency as the time difference between the displacement of the evaluation tool caused by a physical event and the associated virtual event (i.e., the associated movement of the projected virtual object). The physical event is the shooting of a rubber band against the frame, whereas its associated virtual event corresponds to the update of the projection of the virtual point. We repeated the test 15 times and captured the scene with the external camera switched to setting 2. Next, we decomposed the output clips into their frames and visually analyzed them to identify the frame in which the tool moves after the physical event and the frame in which the virtual information consequently updates its location. The number of frames between these two instants was then translated in ms by means of the selected frame rate of 240 fps (setting 2).

2.3.3. User Test

To evaluate how the spatial and temporal registration errors combine whilst the HMP is worn by a user, we adopted the dummy presented in [26]. The dummy consists of three spherical tracking markers attached to a portion of a 3D-printed skullcap derived from a human CT scan. A specific trajectory, traced through Creo Parametric software, was designed to test the system registration accuracy. This trajectory is projected over the dummy in the form of a virtual cutting line that simulates a craniotomy (i.e., the virtual trajectory). On the other hand, we traced a real cutting line on the dummy using a template to obtain a real reference on the correct position of the cutting line (i.e., the real trajectory). The template was designed to exactly match the phantom surface and it was equipped with attachment holes engaging into cylinders mounted on the phantoms. This features ensures a unique and stable mechanical placement of the template on the phantom surface. We asked a user to wear the HMP and employ it to provide an AR view of the cutting line. We captured the scene with the GoPro switched to setting 3, as shown in Figure 2a. We first decomposed the clip into its frames and converted them into Hue-Saturation-Value (HSV) scale. This allowed us to segment the images by adopting a threshold on the H and V channels to extract the real cutting line drawn in blue and the white-colored projected line. Then, the segmented areas were collapsed into a one-pixel line, corresponding to the center line of the real cutting line obtained through skeletonization. The overlay accuracy between the real and projected cutting lines was measured in terms of Hausdorff distance as done in [27], which is defined as the largest distance among all the distances of a point in one line to the closest point of the other line:
d H = m a x { sup x X inf y Y d ( x , y ) , sup y Y inf x X d ( x , y ) , }
Figure 2b, shows the extracted real and virtual centrelines of the real and projected cutting lines, and it highlights the points identifying the Hausdorff distance. This process was repeated for each frame to estimate the total registration error. In addition, the stability of the projected line was evaluated in terms of jitter, which is the deviation from the mean of all the projected lines, and it is used to obtain an estimate of the stability of the virtual content. The latter information was derived by fitting the lines with a 7-order polynomial. The average line was then found by averaging the coefficients among all the captured frames.
Figure 2. User test. (a) Set-up for the user test consisting of the Head-mounted Projector (green), the GoPro Hero 5 (orange), and the processing unit (blue). (b) Hausdorff distance between the projected center line (blue) and the real one (red). Points resulting from the Hausdorff algorithm are marked in black and the line connecting them is the Hausdorff distance between the two curves.
Figure 2. User test. (a) Set-up for the user test consisting of the Head-mounted Projector (green), the GoPro Hero 5 (orange), and the processing unit (blue). (b) Hausdorff distance between the projected center line (blue) and the real one (red). Points resulting from the Hausdorff algorithm are marked in black and the line connecting them is the Hausdorff distance between the two curves.
Sensors 23 03494 g002

3. Results and Discussion

3.1. Tests Results

The spatial registration error within the working area was 11 ± 2 px, which corresponds to 4.6 ± 0.8 arcmin for setting 3 of the external camera. Given the working distance of 475 mm, we estimated a spatial registration error of about 0.6 ± 0.1 mm (Table 2). These spatial registration errors belong to the ranges found in the recent literature for projected AR devices [24,28].
For the temporal registration, the number of frames elapsed between the motion of the evaluation tool and the update of the virtual information is 27 ± 3 on average. Given the acquisition rate of 240 fps (setting 2 in Table 1), this is associated with an average MTP latency value of 113 ± 12 ms (Table 3) This value is far lower than what is expected for [14], where 116 ms corresponds to the latency of the projector alone. On the other hand, the latency obtained in this work amounts is almost twice that reported for the MoSART system [15]. However, in both works, the authors provide no details on the latency evaluation, nor on the tracking or rendering times, thus no insights can be gained into the source of these differences.
We recorded a 25 s clip during the user tests which corresponds to about 6000 frames. Each frame was processed to estimate the centreline of the real and the projected cutting lines. Their Hausdorff distance is plotted in Figure 3a. The average error was found to be 5 ± 1 px corresponding to 6.6 ± 1.1 arcmin and 0.9 ± 0.2 mm (Table 2). The registration error with the HMP worn by a user is higher than the error found for spatial registration alone. This may be due to small variations in the user’s head pose, which combined with the temporal registration error, result in a higher misalignment between real and virtual trajectories. However, the registration error falls within the acceptable range for performing precision manual tasks such as surgery [29]. The jitter with respect to the mean trajectory is plotted in Figure 3b. The RMS of the jitter is 2 px, equal to about 2.5 arcmin and 0.4 mm. This value is close to the one reported for MoSART (0.08 MSE = 0.3 RMS).

3.2. Ergonomics Considerations for Future Development

In terms of perceived user comfort and ergonomics, weight distribution in head-worn displays and projectors is one of the main limiting factors that hinder the widespread adoption of mobile AR devices in the healthcare sector and specifically in image-guided neurosurgery [30,31]. By way of example, the authors of [31] state that, since much of the weight on the Microsoft HoloLens is placed on the user’s forehead and nose, this may force the user to progressively shift his/her gaze downwards over time. This gradual shift in posture could therefore cause neck strain and general discomfort. Similar results were also obtained in [32], a user study where several participants reported back and neck strain as they were forced to move the weight of their head and of the display away from their body. Therefore, and in accordance with the findings of Ito et al. [33], improving the comfort of head-worn displays requires specific considerations regarding weight balance. Therefore, if on the one hand our HMP strictly fulfills the ergonomics requirement in terms of total weight to be <350 g, additional work has to be carried out to improve weight distribution and ergonomics so that the HMP can be worn for a prolonged time.

4. Conclusions

We proposed a device for projected AR, which can be worn by the user to guide the execution of precision manual tasks while preserving mobility in the working area. Compared to HMDs, our HMP offers the advantages of being simultaneously accessible by multiple users, without altering their awareness of the environment and feeling of immersion. We evaluated the performance of our HMP in terms of spatial and temporal registration along with their mutual interaction in a user test. The latency of 113 ms is a positive result, especially considering that the HMP is not equipped with an IMU sensor, nor with predictive slam algorithms or fast-rendering engines such as the ones adopted in the HoloLens. Specifically, the HMP meets the specifications proposed in [34] for HMDs, stating that “latencies shall be not longer than 150 milliseconds”. Future work will be aimed at identifying any critical points to further minimize the latency and evaluate whether it impacts task performance, sickness and discomfort [10,16] even in the case of projected AR. The jitter of 0.4 mm lies within the recommended range for HMDs [19]. It is slightly larger than the value provided for MoSART and twice the value reported for HoloLens 2. Although MoSART offers better performance in terms of latency and jitter, its overall weight and footprint are significantly higher with respect to the device investigated in this study. Specifically, the weight of 1 kg is likely to compromise the usability of the HMP in various applications such in military and medical fields [30,35].
To the best of the authors’ knowledge, our evaluation of spatial and temporal registration, together with their mutual interaction in a user test constitutes the first overall assessment of an HMP for precision manual tasks. We hope that this will provide a reference for further investigation of this and other AR HMPs and HMDs contributing to providing a standard for their performance assessment.

Author Contributions

Conceptualization, V.M. and F.C.; methodology, V.M. and F.C.; software, F.C.; validation, V.M., N.C. and F.C.; hardware, S.C. and R.D., formal analysis, V.M. and F.C.; investigation, V.M. and F.C.; resources, V.F.; data curation, V.M. and F.C.; writing–original draft preparation, V.M. and F.C.; writing—review and editing, V.M. and F.C.; visualization, V.M. and F.C.; supervision, V.F.; project administration; funding acquisition, V.F. and V.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the HORIZON2020 Project VOSTARS (Video-Optical See Through AR surgical System), Project ID: 731974. Call: ICT-29-2016 Photonics KET 2016.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki. Ethical review and approval were waived for this study, based on the nonclinical nature of this study and the anonymous nature of the data collected, which does not allow the tracing of sensitive personal data. The participants were fully informed about the study requirements and had to agree to the data-sharing and privacy policy before participating in the study.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

This work was supported by the Italian Ministry of Education and Research (MIUR) in the framework of the CrossLab project (Departments of Excellence) of the University of Pisa, laboratory of Augmented Reality.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
HMD Head-mounted display
ARAugmented reality
HMPHead-mounted projector
VSTVideo see-through
OSTOptical see-through

References

  1. Fukuda, T.; Orlosky, J.; Kinoshita, T. Head mounted display implementations for use in industrial augmented and virtual reality applications. In Proceedings of the International Display Workshops, Sendai, Japan, 6–8 December 2017; pp. 156–159. [Google Scholar]
  2. Dhiman, H.; Martinez, S.; Paelke, V.; Röcker, C. Head-Mounted Displays in Industrial AR-Applications: Ready for Prime Time? In Proceedings of the HCI in Business, Government, and Organizations, Las Vegas, NV, USA, 15–20 July 2018; Nah, F.F.H., Xiao, B.S., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 67–78. [Google Scholar]
  3. Duncan-Vaidya, E.A.; Stevenson, E.L. The Effectiveness of an Augmented Reality Head-Mounted Display in Learning Skull Anatomy at a Community College. Anat. Sci. Educ. 2021, 14, 221–231. [Google Scholar] [CrossRef] [PubMed]
  4. Morozov, M. Augmented Reality in Military: AR Can Enhance Warfare and Training. Obtido de. 2018. Available online: https://jasoren.com/augmented-reality-military (accessed on 1 February 2023).
  5. Kallberg, J.; Beitelman, M.V.; Mitsuoka, M.V.; Officer, C.W.; Pittman, J.; Boyce, M.W.; Arnold, L.C.T.W. The Tactical Considerations of Augmented and Mixed Reality Implementation. Mil. Rev. 2022, 662, 105. [Google Scholar]
  6. Rahman, R.; Wood, M.; Qian, L.; Price, C.; Johnson, A.; Osgood, G. Head-Mounted Display Use in Surgery: A Systematic Review. Surg. Innov. 2019, 27, 88–100. [Google Scholar] [CrossRef] [PubMed]
  7. Innocente, C.; Piazzolla, P.; Ulrich, L.; Moos, S.; Tornincasa, S.; Vezzetti, E. Mixed Reality-Based Support for Total Hip Arthroplasty Assessment. In Proceedings of the Advances on Mechanics, Design Engineering and Manufacturing IV, Ischia, Italy, 1–3 June 2022; Gerbino, S., Lanzotti, A., Martorelli, M., Mirálbes Buil, R., Rizzi, C., Roucoules, L., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 159–169. [Google Scholar]
  8. Aggarwal, R.; Singhal, A. Augmented Reality and its effect on our life. In Proceedings of the 2019 9th International Conference on Cloud Computing, Data Science Engineering (Confluence), Noida, India, 10–11 January 2019; pp. 510–515. [Google Scholar] [CrossRef]
  9. Aromaa, S.; Väätänen, A.; Aaltonen, I.; Goriachev, V.; Helin, K.; Karjalainen, J. Awareness of the real-world environment when using augmented reality head-mounted display. Appl. Ergon. 2020, 88, 103145. [Google Scholar] [CrossRef] [PubMed]
  10. Stauffert, J.P.; Niebling, F.; Latoschik, M. Latency and Cybersickness: Impact, Causes, and Measures. A Review. Front. Virtual Real. 2020, 1, 582204. [Google Scholar] [CrossRef]
  11. Meyers, M.; Hughes, C.; Fidopiastis, C.; Stanney, K. Long Duration AR Exposure and the Potential for Physiological Effects. MODSIM World 2020, 61, 1–9. [Google Scholar]
  12. Condino, S.; Carbone, M.; Piazza, R.; Ferrari, M.; Ferrari, V. Perceptual Limits of Optical See-Through Visors for Augmented Reality Guidance of Manual Tasks. IEEE Trans. Biomed. Eng. 2019, 67, 411–419. [Google Scholar] [CrossRef] [PubMed]
  13. Nazarova, E.; Sautenkov, O.; Altamirano Cabrera, M.; Tirado, J.; Serpiva, V.; Rakhmatulin, V.; Tsetserukou, D. CobotAR: Interaction with Robots using Omnidirectionally Projected Image and DNN-based Gesture Recognition. In Proceedings of the 2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Melbourne, Australia, 17–20 October 2021; pp. 2590–2595. [Google Scholar] [CrossRef]
  14. Hartmann, J.; Yeh, Y.T.; Vogel, D. AAR: Augmenting a Wearable Augmented Reality Display with an Actuated Head-Mounted Projector. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology, Virtual, 20–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 445–458. [Google Scholar]
  15. Cortes, G.; Marchand, E.; Brincin, G.; Lécuyer, A. MoSART: Mobile Spatial Augmented Reality for 3D Interaction with Tangible Objects. Front. Robot. AI 2018, 5, 93. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Nguyen, T. Low-Latency Mixed Reality Headset. In Low-Latency VR/AR Headset Project from Conix Research Center, Computing on Network Infrastructure for Pervasive Perception, Cognition and Action. 2020. Available online: https://people.eecs.berkeley.edu/~kubitron/courses/cs262a-F19/projects/reports/project14_report.pdf (accessed on 1 February 2023).
  17. Jerald, J.; Whitton, M. Relating Scene-Motion Thresholds to Latency Thresholds for Head-Mounted Displays. In Proceedings of the 2009 IEEE Virtual Reality Conference, Lafayette, LA, USA, 14–18 March 2009; pp. 211–218. [Google Scholar] [CrossRef] [Green Version]
  18. Kress, B.C. Digital optical elements and technologies (EDO19): Applications to AR/VR/MR. In Proceedings of the Digital Optical Technologies 2019, Munich, Germany, 24–26 June 2019; Kress, B.C., Schelkens, P., Eds.; International Society for Optics and Photonics (SPIE): Bellingham, WA, USA, 2019; Volume 11062, pp. 343–355. [Google Scholar] [CrossRef] [Green Version]
  19. Louis, T.; Troccaz, J.; Rochet-Capellan, A.; Bérard, F. Is It Real? Measuring the Effect of Resolution, Latency, Frame Rate and Jitter on the Presence of Virtual Entities. In Proceedings of the 2019 ACM International Conference on Interactive Surfaces and Spaces, Daejeon, Republic of Korea, 10–13 November 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 5–16. [Google Scholar] [CrossRef]
  20. Zheng, F. Spatio-Temporal Registration in Augmented Reality. Ph.D. Thesis, The University of North Carolina at Chapel Hill, Chapel Hill, NC, USA, 2015. [Google Scholar]
  21. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  22. Chang, W.; Wu, C. Plane-based geometric calibration of a projector-camera reconstruction system. In Proceedings of the 2014 10th France-Japan/8th Europe-Asia Congress on Mecatronics (MECATRONICS2014-Tokyo), Tokyo, Japan, 27–29 November 2014; pp. 219–223. [Google Scholar]
  23. Cutolo, F.; Fida, B.; Cattari, N.; Ferrari, V. Software Framework for Customized Augmented Reality Headsets in Medicine. IEEE Access 2020, 8, 706–720. [Google Scholar] [CrossRef]
  24. Mamone, V.; Fonzo, M.D.; Esposito, N.; Ferrari, M.; Ferrari, V. Monitoring Wound Healing With Contactless Measurements and Augmented Reality. IEEE J. Transl. Eng. Health Med. 2020, 8, 1–12. [Google Scholar] [CrossRef] [PubMed]
  25. Cutolo, F.; Cattari, N.; Carbone, M.; D’Amato, R.; Ferrari, V. Device-Agnostic Augmented Reality Rendering Pipeline for AR in Medicine. In Proceedings of the 2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Bari, Italy, 4–8 October 2021; pp. 340–345. [Google Scholar] [CrossRef]
  26. Mamone, V.; Cutolo, F.; Condino, S.; Ferrari, V. Projected Augmented Reality to Guide Manual Precision Tasks: An Alternative to Head Mounted Displays. IEEE Trans. Hum.-Mach. Syst. 2021, 52, 567–577. [Google Scholar] [CrossRef]
  27. Mamone, V.; Ferrari, V.; Condino, S.; Cutolo, F. Projected Augmented Reality to Drive Osteotomy Surgery: Implementation and Comparison With Video See-Through Technology. IEEE Access 2020, 8, 169024–169035. [Google Scholar] [CrossRef]
  28. Gao, Y.; Zhao, Y.; Xie, L.; Zheng, G. A Projector-Based Augmented Reality Navigation System for Computer-Assisted Surgery. Sensors 2021, 21, 2931. [Google Scholar] [CrossRef] [PubMed]
  29. Hussain, R.; Lalande, A.; Guigou, C.; Grayeli, A.B. Contribution of Augmented Reality to Minimally Invasive Computer-Assisted Cranial Base Surgery. IEEE J. Biomed. Health Infor. 2020, 24, 2093–2106. [Google Scholar] [CrossRef] [PubMed]
  30. D’Amato, R.; Cutolo, F.; Badiali, G.; Carbone, M.; Lu, H.; Hogenbirk, H.; Ferrari, V. Key Ergonomics Requirements and Possible Mechanical Solutions for Augmented Reality Head-Mounted Displays in Surgery. Multimodal Technol. Interact. 2022, 6, 15. [Google Scholar] [CrossRef]
  31. Nguyen, N.Q.; Cardinell, J.; Ramjist, J.M.; Androutsos, D.; Yang, V.X. Augmented reality and human factors regarding the neurosurgical operating room workflow. In Proceedings of the Optical Architectures for Displays and Sensing in Augmented, Virtual, 28–31 March 2021; SPIE: Bellingham, WA, USA, 2020; Volume 11310, pp. 119–125. [Google Scholar]
  32. Lin, C.; Andersen, D.; Popescu, V.; Rojas-Muñoz, E.; Cabrera, M.E.; Mullis, B.; Zarzaur, B.; Anderson, K.; Marley, S.; Wachs, J. A First-Person Mentee Second-Person Mentor AR Interface for Surgical Telementoring. In Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Munich, Germany, 16–20 October 2018; pp. 3–8. [Google Scholar] [CrossRef]
  33. Ito, K.; Tada, M.; Ujike, H.; Hyodo, K. Effects of Weight and Balance of Head Mounted Display on Physical Load. In Proceedings of the Virtual, Augmented and Mixed Reality, Multimodal Interaction, Orlando, FL, USA, 26–31 July 2019; Chen, J.Y., Fragomeni, G., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 450–460. [Google Scholar]
  34. Design Criteria Standard: Human Engineering; Department of Defense: Washington, DC, USA, 2020.
  35. Cummings, M.L. Provocations: Technology Impedances to Augmented Cognition. Ergon. Des. 2010, 18, 25–27. [Google Scholar] [CrossRef]
Figure 1. Spatial registration. (a) Set-up for evaluation of spatial registration consisting of the HMP (green) and the GoPro Hero 5 (orange), both placed on a holder for working depth selection, the processing unit (blue), and the spatial registration evaluation tool. (b) Zoom of an image captured by the GoPro Hero 5: pink crosses denote the real reference points, whereas blue crosses denote the projected points.
Figure 1. Spatial registration. (a) Set-up for evaluation of spatial registration consisting of the HMP (green) and the GoPro Hero 5 (orange), both placed on a holder for working depth selection, the processing unit (blue), and the spatial registration evaluation tool. (b) Zoom of an image captured by the GoPro Hero 5: pink crosses denote the real reference points, whereas blue crosses denote the projected points.
Sensors 23 03494 g001
Figure 3. Results of the user test. (a) Hausdorff distances between the real trajectories and their corresponding projected trajectories. (b) Jitter of the projected trajectories.
Figure 3. Results of the user test. (a) Hausdorff distances between the real trajectories and their corresponding projected trajectories. (b) Jitter of the projected trajectories.
Sensors 23 03494 g003
Table 1. GoPro Hero 5 black: settings and specifications.
Table 1. GoPro Hero 5 black: settings and specifications.
SettingTestResolutionFrame-RateAngular ResolutionWorking Depth
1Spatial registration3000 × 4000-0.40 arcmin/px350–600 mm
2Temporal registration1280 × 720240 fps1.23 arcmin/px475 mm
3User test1280 × 720240 fps1.23 arcmin/px475 mm
Table 2. Spatial registration errors.
Table 2. Spatial registration errors.
SettingTestOn Image Registration
Error (px)
Visual Angle Registration
Error (arcmin)
Absolute Registration
Error (mm)
1Spatial registration 11 ± 2 4.6 ± 0.8 0.6 ± 0.1
3User test 5 ± 1 6.6 ± 1 0.9 ± 0.2
Table 3. Temporal registration errors.
Table 3. Temporal registration errors.
SettingTestMotion-to-Photon Latency (ms)
2Temporal registration 113 ± 12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mamone, V.; Ferrari, V.; D’Amato, R.; Condino, S.; Cattari, N.; Cutolo, F. Head-Mounted Projector for Manual Precision Tasks: Performance Assessment. Sensors 2023, 23, 3494. https://doi.org/10.3390/s23073494

AMA Style

Mamone V, Ferrari V, D’Amato R, Condino S, Cattari N, Cutolo F. Head-Mounted Projector for Manual Precision Tasks: Performance Assessment. Sensors. 2023; 23(7):3494. https://doi.org/10.3390/s23073494

Chicago/Turabian Style

Mamone, Virginia, Vincenzo Ferrari, Renzo D’Amato, Sara Condino, Nadia Cattari, and Fabrizio Cutolo. 2023. "Head-Mounted Projector for Manual Precision Tasks: Performance Assessment" Sensors 23, no. 7: 3494. https://doi.org/10.3390/s23073494

APA Style

Mamone, V., Ferrari, V., D’Amato, R., Condino, S., Cattari, N., & Cutolo, F. (2023). Head-Mounted Projector for Manual Precision Tasks: Performance Assessment. Sensors, 23(7), 3494. https://doi.org/10.3390/s23073494

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop