Next Article in Journal
A Lever Coupling Mechanism in Dual-Mass Micro-Gyroscopes for Improving the Shock Resistance along the Driving Direction
Previous Article in Journal
Humidity Sensor Based on Bragg Gratings Developed on the End Facet of an Optical Fiber by Sputtering of One Single Material
Previous Article in Special Issue
Model-Based Estimation of Ankle Joint Stiffness
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Instrumented Glove to Assess Manual Dexterity in Simulation-Based Neurosurgical Education

1
Bioinstrumentation and Clinical Engineering Research Group—GIBIC, Bioengineering Department, Engineering Faculty, Universidad de Antioquia UdeA, Calle 70 No. 52-21, Medellín 050010, Colombia
2
LAAS-CNRS, Université de Toulouse, CNRS, Toulouse 31400, France
3
ISIFC, Université de Franche-Comté, Besançon 25000, France
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(5), 988; https://doi.org/10.3390/s17050988
Submission received: 31 October 2016 / Revised: 1 April 2017 / Accepted: 7 April 2017 / Published: 29 April 2017
(This article belongs to the Special Issue Wearable Biomedical Sensors)

Abstract

:
The traditional neurosurgical apprenticeship scheme includes the assessment of trainee’s manual skills carried out by experienced surgeons. However, the introduction of surgical simulation technology presents a new paradigm where residents can refine surgical techniques on a simulator before putting them into practice in real patients. Unfortunately, in this new scheme, an experienced surgeon will not always be available to evaluate trainee’s performance. For this reason, it is necessary to develop automatic mechanisms to estimate metrics for assessing manual dexterity in a quantitative way. Authors have proposed some hardware-software approaches to evaluate manual dexterity on surgical simulators. This paper presents IGlove, a wearable device that uses inertial sensors embedded on an elastic glove to capture hand movements. Metrics to assess manual dexterity are estimated from sensors signals using data processing and information analysis algorithms. It has been designed to be used with a neurosurgical simulator called Daubara NS Trainer, but can be easily adapted to another benchtop- and manikin-based medical simulators. The system was tested with a sample of 14 volunteers who performed a test that was designed to simultaneously evaluate their fine motor skills and the IGlove’s functionalities. Metrics obtained by each of the participants are presented as results in this work; it is also shown how these metrics are used to automatically evaluate the level of manual dexterity of each volunteer.

1. Introduction

Fine Motor Skills (FMSs) or dexterity is the coordination of small muscles, in movements—involving the synchronization of wrists, hands, and fingers—with the eyes. FMSs are the set of abilities that humans use when performing complex actions such as picking up small objects, writing, painting or sewing [1,2]. FMSs appear at the moment of birth; in childhood, motor skills continue to evolve, particularly during school age. Refinement of motor skills in adults is strongly influenced by the daily activities they carry out. Motor skills decrease significantly as a result of physical and cognitive deteriorations caused by aging [3].
A high degree of FMS is an essential requirement for anyone who wants to become a surgeon; this is much more significant in medical specialties such as neurosurgery, as small errors may imply severe consequences for the patient. FMS is related to the deftness with which surgical instruments are manipulated, the gentleness while handling tissues, the time and amount of movement required to perform each task, the degree of precision for reaching target areas, and the amount of force applied [4,5]. Relatively few studies have been conducted to find out key aspects to be evaluated in future neurosurgeons [6,7]. However, there is an agreement around the relevance of evaluating hand dexterity, because it allows a more efficient technique and greater ability to achieve success in surgical procedures [8,9]. Assessment of this parameter has often been subject to criteria of instructors, or at its best, to the review by a panel of experienced surgeons who examine video recordings from operations performed by residents [5]. The use of surgical simulators as training tools can be considered as a complement to the traditional master-apprenticeship approach. In those mock procedures, trainees are asked to repeat simulation exercises on “simulated patients”, until they have learned the required skills to deal with “real patients” [10]. As a consequence of the situations mentioned above, future surgeons will improve their abilities insofar as they spend time in front of a simulator (most of the time, there will be no experts nearby to help). Therefore it is necessary to implement tools to enable an efficient assessment of manual dexterity in surgical simulations.
Formerly, simulators were focused only on training in specific tasks; this technology has now evolved. Thus new advanced simulators are not only training systems, but they can be assessment tools at the same time [11]. For their use in FMS assessment, surgical simulators should include a combination of hardware and software that enables to measure and process information from variables such as hand movements, applied forces, trajectories and timing [5,9,12,13,14,15,16]. These features are now available in most advanced devices like LapSim and LapVR, for laparoscopic surgery or NeuroVR, for neurosurgical simulation [17,18,19]. Those simulators use haptic peripherals to track hand movements and also recreate the sense of touch on Virtual Reality (VR) environments [18,19]. However, most haptic meaning-based applications are expensive to develop and requires a big research team, since they often entail significant engineering challenges and the use of advanced computational models [9,20]. Modern VR simulators still need to be improved, to match real models in aspects like tissue behavior, tactile sensation, and anatomical appearance [21]. Some of these limitations can be partly attributed to the small ecosystem of haptic peripherals available, so, the range of possible applications for the simulation of surgical instruments is often limited to simple pen-like or scissors-like devices [22,23,24]. Alternatively, traditional bench top and manikin-based simulators use physical models that emulate anatomical structures. Those models are quite realistic, giving the trainee the impression of working with “real things”. Also, they are made with materials that resemble real tissues, making touch sensations lifelike [11]. Although this technology has been used in medical education for more than fifty years, it still lacks tools for hand tracking or manual dexterity assessment.
To improve the capabilities of traditional simulators, many custom hardware/software designs useful for hand motion analysis and surgical dexterity assessment have been developed; they can be categorized to non-glove and glove-types. Datta et al. [25,26] used an electromagnetic tracker as a motion analysis system that measures the number of hand movements made and the time taken to complete each task and found that hand motion analysis may be an effective objective measure of dexterity in open surgical simulation. Ross et al. developed a semi-automated catheter-tracking software from motion analysis of fluoroscopic videos, and they found that it is a valuable tool for the objective assessment of endovascular skills [27]. Solutions that used image processing are based on cameras which use the visible or infrared spectrum [28]; some of them, take direct pictures of the hand [29,30], while others rely on visual detection of optical markers in the hand [31]. In general, they are accurate but require line-of-sight (a straight path without obstacles between cameras and tracked objects), and as an additional drawback, they are affected by lighting, so they are only useful in environments with well-controlled conditions. Zappella et al., Gray et al., and Uemura et al. used digital video systems to capture hand and instruments movements of surgical trainees, information is processed and used to perform surgical gesture classification and assess laparoscopic skills objectively [32,33,34]. Harada et al. designed instrumented tweezers with the ability to measure several parameters of instrument manipulation during anastomosis simulations, hardware include an Infrared optical motion tracking system, IMUs, and strain gauges [35]. Hammond et al. designed a similar device, using an electromagnetic tracker, and also they presented a novel sensor technique to print metallic strain gauges on the surface of surgical instruments [36,37]. Zihni et al., and Overby and Watson used IMU sensors to register hand movements and instrument trajectories. Acquired data were numerically manipulated using different digital signal processing approaches to produce metrics to assess FMS [38,39,40]. Finally, electromyographic (EMG) signals have been used for hand movement processing, they can appropriately recognize basic hand gestures, but they are useless for fine finger movements analysis [41]. Additionally, EMG has the disadvantage of requiring the installation of surface electrodes [42], which usually are uncomfortable during training, and require some level of expertise to make sure that electrodes are correctly installed.
Glove solutions or Instrumented Gloves (IGs) are wearable systems that have been used in hand movement analysis for years; this is the reason why there are many IGs available on the market [43,44,45]. Although those IGs are now a relatively mature technology, they are in general too bulky to be used comfortably in medical simulation. Many of these IGs that can be found in the literature are designed as tools for the treatment of people who has Parkinson’s disease or suffered a stroke [46,47,48]. IGs equipped with optical encoders, or fiber-optic nerves have been developed [49,50]. This technology made it possible to produce IGs that fit different hand sizes, and they also bring a good movement freedom, but they can only record the movement of fingers with respect to the hand (they cannot provide information about hand movement). Something similar happens with IG based on bend sensors, they commonly use force sensitive resistors used as flex sensors to measure movements of each finger [51,52]. Therefore, there is no data associated with the hand itself. IGs based on electromagnetic sensors can detect the movements of fingers and hand [15,53]. However, this kind of systems requires robust computational processing to reduce the susceptibility to electromagnetic interference generated by other electric devices and metal objects [54]. Finally, IGs based on Inertial Movement Units (IMU) have recently emerged as a powerful alternative [55,56]. IMUs are quite small, as they can be unobtrusively embedded on wearable designs [57], this avoids using bulky and loose parts on IGs. The benefits of these IGs are plentiful; they can monitor the movement of the hands and fingers, they do not require line of sight, and they are not affected by electromagnetic interference.
The aim of this paper is to present the IGlove system, a wearable device which is composed of an IG that uses IMUs to capture hand movements and a set of software applications that process IMUs signals to estimate some of the metrics that are commonly used by modern surgical simulators to assess trainee’s manual dexterity. IGlove was developed in cooperation between GIBIC group in Colombia and S4M group in France [58,59]. The main objective of this paper is to show how to process acquired data to obtain hand dexterity metrics that are suitable to assess FMS on surgical simulation environments automatically. IGlove was originally conceived to be integrated to Daubara NS Trainer (a neurosurgical simulator previously developed by GIBIC group), but it could be easily integrated to another manikin- or benchtop-based medical simulators, or even in other applications where hand motion analyses are required.
The rest of the paper is divided as follows: Section 2.1 provides a detailed description of the IGlove’s hardware architecture. Section 2.2 shows the details regarding the design of a test for assessing FMS of a group of volunteers using IGlove, in which each volunteer was filmed while he was doing the test. Section 2.3 shows how FMSs were evaluated “manually”, according to the criteria of human evaluators. Section 2.4 describes algorithms utilized to obtain manual dexterity metrics. Section 2.5 makes a dimension reduction analysis, in order to assess the real necessity of using all the data acquired by each sensor. If it is found that the analysis can be carried out using fewer signals or sensors, the complexity of the problem may be significantly reduced. Section 2.6 presents some statistical tools to classify volunteers according to their level of FMS based on the previously-mentioned metrics. Section 3 shows the results of IGlove’s test, which were obtained from the methodological design proposed in Section 2. Section 4 presents a discussion of the results and their applications in neurosurgical education. Finally, conclusions and future trends are presented in Section 5.

2. Materials and Methods

2.1. IGlove Description

The IGlove hardware consists of two main parts: the Glove and the Hub (Figure 1a). The Hub’s architecture is based on an electronic board that features a Cypress PSoC4 CY8C4245AXI-483 as the core of this system; this ARM Cortex M0 microcontroller is responsible for setting up an SPI interface with IMU sensors on Glove [60]. The Hub features a power supply based on an LDO regulator that converts 5 V from USB port to 3.3 V to feed the circuit. Acquired data is transmitted to the Daubara NS Trainer using a USB to serial UART bridge; processing algorithms will be discussed in Section 2.4.
The electronics are housed in an ergonomic custom-designed enclosure. The Hub is secured to user’s arm by using an adjustable strap (see Figure 1b). For the glove fabrication, several fabrics were tested, a thin and elastic fabric was chosen to ensure users comfort during exercises. To warranty hygiene and cleanliness gloves are washable. Also, there is a glove for each user. The glove comes in two sizes, large and small (especially useful for people with small hands). IGlove’s electronics do not have direct electrical contact with skin and has been manufactured with materials that are inert to the skin. Therefore, this wearable design does not represent drawbacks of biocompatibility or electrical safety. IMU Sensors are installed in specially designed compartments (small pockets) on the glove, where sensors can be precisely adjusted, as shown in Figure 2a. IGlove’s sensors can be easily removed from the glove; then they can be used with the next user. This feature allows customizing the sensor distribution to perform different types of studies, just by placing sensors in various parts of the hand. Current Hub configuration permits the connection of up to eight sensors, but it can be easily extended.
In order to determine the appropriate number of sensors to be placed in the glove, a group of neurosurgeons and residents of neurosurgery was observed while handling some surgical instruments like biopsy needles, dissectors, and ultrasonic aspirator. It was found that during manipulation of those instruments, specialists preponderantly used only three fingers, thumb, index and middle finger; ring finger and little finger were rarely used. For these reasons, it was decided to embed four sensors on IGlove, on distal phalanges of fingers commonly used for instrument handling (thumb, index, and middle) and on the back of the hand. However, the volume of data from those four sensors (which generate 36 signals in total), translates into a huge challenge regarding analysis and processing of information. On Section 2.5 a procedure which identifies sensors and signals that are more relevant for FMS assessment is presented. Thus, the redundant and irrelevant information can be ignored, significantly reducing the complexity of the problem.
A sample of each sensor is shown in Figure 2b. It is based on MPU-9250 from InvenSense, which is a multi-chip module consisting of two dies integrated into a single package. One die houses the 3-axis gyroscope and the 3-axis accelerometer. The other die houses a 3-axis magnetometer. Hence, the MPU-9250 is a truly 9-Degrees of Freedom (DOF) motion tracking device in a 3 × 3 × 1 mm QFN package [61]. A little circuit board was designed to mount the IMU, due to the severe size constraints imposed by the small sensor compartments. Special considerations regarding connectors and communication lines that connect sensors to hub were taken. As cables will be draped over the user, those cables should slide freely without tangling with instruments used in the simulation. Cables must also bend, twist, and flex without impeding hand movements. A little 8-pin connector was used to bring power lines (GND and +3.3 V) and communications (four-line SPI and interrupt) from the hub. For those purposes, 0.25” pitch Wired Polarized Nano Connectors from Omnetics were used, they are small enough and incorporate exceptionally thin and flexible cables [62].

2.2. Test Design

A total of 14 volunteers were recruited to perform a test for evaluating IGlove’s functionalities. Ten trainees were male, and four were females, age range between 23 to 32 years old, all of them were right-handed, and none of them had previous contact with medical simulators or psychometric tests for measuring hand dexterity. They manifested had not consumed drugs or alcohol in the past two days, and they are not suffering from any illness or condition that would prevent them from doing the test. All volunteers were informed in detail about the study, and all of them signed an informed consent for participating in this study. The test consisted of three exercises to be performed while they wore the IGlove. Between exercises, there were periods of rest, during these periods the participant was asked to put his/her hand on the table in a neutral position. Rest periods were included with a dual purpose, first, give a break to the volunteer, preventing fatigue to affect his/her performance, and secondly, as it will be shown below, to facilitate exercises segmentation.
The first exercise prompted the participants to run the group of hand gestures shown in Figure 3; after each gesture, the participant must put his/her hand in the rest position for a few seconds. These gestures were designed to verify the ability of IGlove to register fine finger movements; rest periods were used to simplify movement recognition process. In the second exercise, the participant had to run the same sequence, but this time, he/she performed the entire sequence without putting his/her hand in the rest position until finished. This exercise was designed to assess the ability of the system to recognize movements without the help offered by rest periods.
The third exercise of the test was focused on FMS assessment. It was designed based on traditional Neuropsychological Tests (NT), which have been concerned with the objective measurement of skills in psychometrics. PegBoard Tests (PBT), like Grooved PegBoard and Purdue PegBoard, are NTs that have been specifically designed as tools for FMS assessment, but commonly they are only used to evaluate motor or cognitive impairment in patients and for personnel selection purposes [63,64]. However, some authors have explored the possibility of using this type of NTs to assess dexterity in surgeons [7]. Researchers have found that there is a significant relationship between scores on NTs that evaluate FMS and the performance of trainees in surgical procedures [65,66,67]. Moreover, Kyle et al. [68] suggest that NTs may be used to identify those novice trainees with lower FMS scores, who might require additional support for their learning of some specific technical tasks. In designing the third exercise, it was also taken into account that most of the laparoscopic simulators include PBT-like modules called Peg Transfer Boards [69]. Those PBT-like modules are used to assess and improve FMS by requiring the surgical trainee to transfer small objects between designated pegs using laparoscopic instruments [67,70]. The third exercise of the test was also designed as a PBT-like module. A star-shaped pegboard was used, featuring many pegs and three deposits for small multicolored Plastic Bushings (PBs), as shown in Figure 4a. Volunteers were given the task of moving six blue PBs from deposits to blue pegs, and in the same way, six red PBs to the vertices of the star (six red pegs), employing, for this purpose small tweezers. Figure 4b presents a volunteer of the test while running this exercise. This kind of exercises was selected due to its ease of implementation, and its extensive use as a tool for assessment FMS which ensures its validity [70]. Trainees did the test only one time because the idea was to use the third exercise as a psychometric test to measure their current level of dexterity. That is the reason why they were asked if they had ever made a test like this one before.

2.3. Video Review of Test

All volunteers were filmed while performing the test, much more attention was given to the third exercise, as it was specifically designed to assess FMS. Videos were subsequently used to achieve two objectives, first, to ensure that exercises were correctly performed, and second, to perform an assessment of the test through a careful examination of video records (human-aided assessment). As described above, the third exercise involves to take a PB from one deposit and place it into the corresponding peg; this action was repeated twelve times. Each repetition of the exercise was segmented into two basic periods of time: firstly, Task1, which starts when a PB is picked up from one deposit using the tweezers and placed onto the appropriate peg (it is the action of transporting the PB). Secondly, Task2, which consists of the remaining shares that must be done just before picking up the next PB. In some cases, Task2 may not only represent the amount of time required to return to pick up another PB, since participants often made mistakes, performed unnecessary movements, and wasted time in trying to find out the easiest PB to pick up from deposits. Time wasted on making these actions was added to Task2 as a penalty mechanism. Some of the actions considered as errors include: drop a PB, misplace a PB or place it in a wrong place, try to use the other hand, perform any sudden or unnecessary movements, and fail to pick up a PB from a deposit.
At the end of the exercise TSpan, TotalTask1, and Errors parameters were calculated. TSpan indicates how long was the exercise (sum of Task1s and Task2s), TotalTask1 is calculated from the from the sum of time invested by each participant for transporting PBs (sum of Task1), Errors is the number of errors during the exercise according to the criteria explained in the previous paragraph. Under these assumptions, people with better FMS should get lower values of TSpan, TotalTask1, and Errors as they can do the exercises quickly and make fewer mistakes. Parameter TRatio is also calculated as TRatio = TSpan/TotalTask1; it is an indicator of efficiency in the exercise; participants who did the exercise efficiently, obtained low TRatio values, as they employed most of the time to transport PBs. Conversely, actions such as taking too long to pick up a PB, make mistakes, correct previous errors or make unnecessary movements, cause an increase in TRatio, indicating lower performance.
A clustering analysis was made based on the information gathered from video review; it was used to classify individuals according to the demonstrated abilities in the third exercise. The standardized Euclidean distance was used as a metric to measure dissimilarity on data. Ward’s linkage was used for clustering.

2.4. IGlove Data Processing

The data processing scheme to obtain metrics from each sensor is briefly depicted in Figure 5. An initial pre-calibration process was carried out to eliminate offset errors on sensor measurements. Section 2.4.1, Section 2.4.2 and Section 2.4.3 give detailed descriptions of signal processing and information analysis performed to obtain manual dexterity metrics. It is worth to mention that the system can calculate absolute orientation angles using sensor fusion algorithms [71]. However, this feature is not shown in Figure 5, since these angles were not utilized in the analysis described in this paper (as shown in Section 2.5 and Section 3.3).

2.4.1. Exercises Segmentation (Estimating Metric Duration)

The first analysis was focused on developing a mechanism to automatically determine the duration of exercises, i.e., an algorithm for detecting the beginning ( t i n i ) and the end ( t e n d ) of each of the exercises. To facilitate this process, volunteers were asked to put right hand in a resting position on the table for several seconds at the end of each of the exercises. Thus, time periods of activity corresponded to those where sensors detected a significant amount of movement [72].

2.4.2. Movements Segmentation (Estimating Metric Movements)

The second analysis was focused on estimating the number of movements carried out within exercises. For this purpose, the magnitude of angular speed ( S ω ) was calculated from the X, Y and Z gyroscopic signals using Equation (1):
S ω = ω x 2 + ω y 2 + ω z 2 2
S ω signal was used to discriminate individual hand movements, as it was observed that peaks on S ω signal occurs each time the hand movement changes (e.g., a change between Task_1 and Task_2) [73]. Therefore, those peaks were used to segment the movements in epochs. Figure 6a shows a segment of a typical S ω signal in the third exercise. The first peak corresponds to the moment where a PB is collected. The interval linking this first peak with the second corresponds to the transport of a PB to the board. The second peak corresponds to the time in which the PB is put on the peg. The interval linking the second peak with the third corresponds to the return to the reservoir. Finally, the last peak corresponds to the action of taking the next PB. It was also found that only some peaks on S ω were well shaped as those shown in Figure 6a. Often, it was common to find peaks as shown in Figure 6b. Many of them with a different shape (high frequency), others too small to be considered a consequence of voluntary movements, and were probably caused by small motion artifacts or even can be attributed to errors in measurement. In some cases, sets of peaks (groups of peaks very close to each other) appear while the volunteer performed the movement. This kind of events are related with variations in the rotation speed while performing a hand movement, a particular way to do the exercise.
An algorithm to eliminate artifact spikes from S ω signal was developed. Firstly, | S ω ^ | the absolute value of the Hilbert’s transform of S ω is calculated. Secondly, resulting signal is smoothed by a FIR filter, using a sliding Hanning window filter (window length = 200 ms). Finally, a peak detection algorithm removes all those cases of peaks that cannot be attributed to hand movements (as it was shown in Figure 6b), the algorithm skips those spikes that are too small to be considered voluntary movements and treats peaks very close to each other as a single movement. Movements performed during an exercise are represented as time intervals between each pair of remaining peaks. This motion sensing technique can also detect erroneous movements. If the amount of movements required to complete an exercise is known, then, additional peaks will evidence the presence of errors or unnecessary movements. Even if the “ideal” amount of movements is unknown, it is also possible to make comparisons with the amount of movements required by a very skilled individual. On surgical simulators, a parameter called movement economy is used; it is calculated as a ratio between the average of movements required by skilled surgeons (taken as a gold standard) and the average of movements performed by a trainee [65].

2.4.3. Kinematic Analysis

Surgical simulators use metrics based on the angular and spatial displacement; it has been found that the most skilled users and experienced surgeons perform much shorter paths in simulations [74]. From the numerical integration of the signals from the accelerometers and gyroscopes, estimation of the spatial and angular displacements can be obtained using IGlove system.
The area under the curve of S ω was estimated to obtain a measurement of angular displacement during an exercise ( D i s p θ ). S ω Signal was previously detrended to reduce gyroscope error issues:
D i s p θ = t i n i t e n d   S ω d t
The magnitude of acceleration signal S a was calculated from the X, Y and Z components of acceleration and subtracting the magnitude of the gravitational acceleration (g):
S a = a x 2 + a y 2 + a z 2 2 g
V a ( t ) which is the function of speed magnitude can be obtained by solving the Equation (4):
d V a ( t ) d t = S a ( t )
To solve Equation (4) Runge-Kutta method was used. Numerical integration was performed for each hand movement, i.e., between each pair of consecutive peaks of S ω . The initial condition for V a on each movement was the speed calculated at the end of the previous movement; the first initial condition V a ( t i n i ) = 0 , as the hand was in a rest position at the beginning of the exercise. Performing speed calculations for each of the movements and not for the whole exercise reflects hand dynamics better and contributes to reducing offset errors of numerical integration.
Calculation of total absolute displacement during an exercise ( D i s p ), can be numerically calculated as the area under the curve of the absolute value of the signal V a ( t ) , the numerical integration was performed using the trapezoidal technique. The absolute value is necessary, otherwise, the value of the total displacement would be zero:
D i s p = t i n i t e n d | V a ( t ) | d t

2.5. Dimension Reduction

Dimension reduction analyses were carried out to evaluate the real necessity of using all the data acquired by each sensor on IGlove, with the objective of determining if the hand motion assessment can be carried out using fewer signals and sensors, if this can be done, the complexity of the problem may be significantly reduced. The correlation coefficient ( r ) was calculated between signals from the four sensors, with the premise that if a high r between two sensors is found, one of them is redundant (so, one of them is not considered for the analysis). Additionally, another analysis was conducted, focusing on whether or not it is necessary to integrate the signal from 3-Axial magnetometers. Although, those sensors are not easy to use, since they require calibration to deal with Hard- and Soft-Iron interference [75], Earth’s magnetic field measurements can be used together with data from accelerometers and gyroscopes to determine the absolute orientation angles ( ϕ , θ and ψ ), by using fusion algorithms [71]. Approximations of angular speeds on each axis can be calculated by approximate derivatives:
ω ϕ = d ϕ d t   ω θ = d θ d t   ω ψ = d ψ d t
Similarly, to Equation (1), the magnitude of angular speed calculated from fusion algorithms ( S F ω ) can be calculated using Equation (7):
S F ω = ω ϕ 2 + ω θ 2 + ω ψ 2 2
A comparison between results obtained with Equation (1) (that used data measured by accelerometers) and Equation (7) (that used data estimated from fusion algorithm) was performed to see whether these magnetometer data were relevant or not.

2.6. Statistical Analysis

From the results above, a clustering of all subjects who participated as volunteers in the test was performed. Metrics Duration, Movements, D i s p θ , and D i s p , were used as variables for this purpose. Once again, the standardized Euclidean distance was used to measure dissimilarity and Ward’s, algorithm was used for clustering. Results of this automatic classification were compared with the results previously obtained from video review. Additionally, a nonparametric test (Kruskal-Wallis, an alternative to the ANOVA test), was used to assess the validity of the classification obtained from clustering. The selection of the nonparametric test was based on the small size of our sample and the stochastic heterogeneity of the samples. The purpose of this test was to evaluate that each one of the variables has significant different distributions across both groups (various dexterity levels). The significance level of the Kruskal-Wallis test was set to 1%, and the comparison was performed using the Bonferroni method for correction.

3. Results

3.1. Results from Video Review

Through the video review, it was found, as expected, that the first two exercises were not a challenge for the participants, all of them did exercises one and two without any significant problem. By contrast, the third exercise showed substantial differences between subjects. Table 1 shows results of measurements obtained from video examination. Each row in the table represents the outcomes of a volunteer who did the test. A cluster analysis was performed using two combinations of TSpan, TotalTask1, TRatio, and Errors (see Table 1); resulting dendrograms are shown in Figure 7. Three groups can be clearly differentiated, blue (volunteers 1, 4, 5, 9, 12 and 13), red (volunteers 7, 8, 11 and 14) and green (volunteers 2, 3, 6 and 10); volunteer 10 can be considered either green or red. Table 2 compares averages and standard deviations for TSpan, TotalTask1, TRatio, and Errors on each group. As can be seen, participants in the blue group had “better numbers”, as they in general, take less time to finish the exercise, had lower TRatio values and made fewer errors. Conversely, the red group had “poorer performance”, as they in general, take long times to finish the exercise, had high TRatio values and made many errors. The green group, according to their times, TRatio values, quantities of errors, can be considered an “average group”. According to this, volunteers 1, 4, 5, 9, 12 and 13, have higher dexterity; volunteers 2, 3, 6 and 10, have an average level of dexterity; and volunteers 7, 8, 11 and 14, have a low degree of dexterity.

3.2. Results from IGlove Data Processing

3.2.1. Exercises Segmentation

Figure 8 shows the results of the segmentation algorithm for each of the exercises. The upper graph corresponds to the signal of a gyroscope from one of the participants (Sensor 1, gyroscope on Z-axis),   S G 1 z acquired throughout the whole test (the signal from any of the 12 gyroscopes on IGlove could be used). An envelope of this signal   | S G 1 z ^ | , was calculated using the absolute value of the Hilbert transform of S G 1 z shown in the second graph (in green), it is a smooth curve that outlines the peaks of S G 1 z and eliminates most of zero crossings. The envelope signal was passed through a comparator. The comparator output is 100% when | S G 1 z ^ | is higher than a Threshold (the threshold is 15 degrees per second, which is high enough to avoid false activation). If | S G 1 z ^ | is lower than Threshold, the output is set to 0%, indicating a rest period. Small false-positives can be found from time to time in the output signal, mainly caused by small involuntary movements. Also, the output may drop to 0% for short intervals when zero-crossings were found on | S G 1 z ^ | signal. To overcome these small errors, all peaks in the output of the comparator signal with a duration time of one second or less were eliminated; similarly, all small timeslots (holes) with output at 0% were filled. Table 3 shows the estimated duration time of each exercise.

3.2.2. Movements Segmentation

Figure 9 Shows results for movement detection in exercise 3 for subjects 12 and 2. Subject 12 was the volunteer with higher hand skills according to video reviewing, and the only one with no errors during this exercise. All peaks detected were marked with green triangles. As previously stated, each peak represents a change in voluntary movements, e.g., in Figure 9a; the second peak marks the instant when the first PB is placed on a peg (instant between Task1 and Task2). Subject 12 required only 25 movements to complete the exercise, other participants who committed errors (like subject 2), required a higher number of movements. Table 4 summarizes results of automatic identification of movements.

3.2.3. Kinematic Analysis Results

A signal of the magnitude of the acceleration S a was calculated using Equation (3) from data acquired by accelerometers. By numerical integration of S a , the signal of the magnitude of the velocity V a was calculated using Equation (4). The absolute displacement, D i s p , i.e., the whole amount of movement made by the volunteer during an exercise is obtained by calculating the area under the curve using Equation (5). An example of this data processing for participant 12 while playing exercise 3 is shown in Figure 10. The total angular displacement during the third exercise was calculated from the signal S ω using Equation (2). Disp, is a measurement of “how much the volunteer moved his/her hand during an exercise”. D i s p θ is a measurement of “how much the volunteer rotated his/her hand during an exercise?” Table 5 shows results of spatial displacement and Table 6 shows results of angular displacement for each volunteer for the three exercises of the IGlove test.

3.3. Dimension Reduction Results

IGlove test was divided into three exercises, an exercise can be divided into all movements that a participant made to complete it, and movements were segmented based on S ω Signal peaks. Table 7 shows the results of a study that focused on the estimation of how similar were movements registered by each of the sensors during the third exercise. To do this, r between S ω signal segments that correspond to each movement was calculated. Subsequently, the average of all these r and their standard deviation were calculated. r was calculated between sensor 1, located in hand and the other sensors on each finger. Also, the correlation between sensors 2 and 3 is shown to evaluate the activity of thumb-index combination, which is used on tweezers manipulation. Table 8 shows results from the same analysis but for exercises 1 and 2. Table 7 and Table 8, shows r calculated between the sensor 1 (in the hand), and the sensors 2, 3 and 4, located on the fingers.
The results for exercises 1 and 3 revealed that in those exercises where r is very high, most of the kinematic behavior of the whole hand could be deduced from data acquired on one of the sensors since the information provided by other sensors is redundant regarding a simple measure of cross-correlation. Therefore, analyses regarding FMS were performed using only signals acquired by sensor 1 (extrapolation to other sensors is possible).
Sensor fusion was used to evaluate the utility of including magnetometer signals in the analysis of hand movement. For this purpose, absolute orientation angles ϕ , θ and ψ were calculated based on information from accelerometers, gyroscopes and magnetometers on sensor 1. Angular speeds ω ϕ ,   ω θ and ω ψ were calculated using Equation (6). The magnitude of angular velocity S F ω was calculated using Equation (7). Similarly, the signal of the magnitude of angular velocity S ω was also calculated using Equation (1) from gyroscopes’ raw data. As an example, Figure 11a shows S F ω signal and Figure 11b the S ω signal from data acquired from volunteer 12. As it can be seen, signals were quite similar, the only significant difference is that the first graph is a little bit delayed with respect to the second one (due to a small delay introduced by the fusion algorithm). In this example r between S F ω and S ω , r = 0.89, (so they are very similar); these high correlation can also be found in the other exercises, and in all the exercises carried out by the other volunteers (r > 0.82 in all cases). So, it was found that S ω gives basically the same information that S F ω . So, no significant advantages are obtained from sensor fusion; and therefore, signals from magnetometers were not considered. However, in those situations where finger movements have a greater role, information regarding absolute rotation would be useful, this fact will be discussed further on Section 4.

3.4. Statistical Analysis Results

A clustering analysis was made based on the information gathered from IGlove during exercise 3, resulting dendrograms are shown in Figure 12. As a remarkable fact, it can be seen that the algorithm was able to identify the same volunteers with higher FMS that were previously identified in Section 3.1.
Participants were clustered into two groups:
D e x t e r i t y   T y p e   1 = { 1 , 4 , 5 , 9 , 12 , 13 } D e x t e r i t y   T y p e   2 = { 2 , 3 , 6 , 7 , 8 , 10 , 11 , 14 }
Table 9 shows the average and standard deviation of Duration, Movements, Disp and D i s p θ , for both groups, as they can be seen, Duration is the variable that better describes the differences between groups, followed by Movements and Disp. It must be noticed that Dexterity_Type_1 group corresponds to those volunteers with higher FMS. This result coincides with the result previously found in Section 3.1.
Also, results of statistical validation using Kruskal-Wallis test are presented in Table 10. As can be seen, there are statistical differences between groups in exercise 3, detected by cluster analysis with variables Duration, Movements, Disp and D i s p θ . Concerning exercises 1 and 2, statistical result of the test showed that they cannot be used to assess dexterity.

4. Discussion

A total of 14 volunteers were recruited to test the IGlove; this quantity may look relatively small, but it should be noted that this population size is consistent with the limited number of residents in most neurosurgery programs [76]. In fact, the neurosurgery program at Universidad de Antioquia (the users of Daubara NS Trainer), admits only three residents per year; there are currently 15 active residents, representing around 15% of all neurosurgical residents in the whole country. To increase the number of potential users of IGlove, and therefore the amount of possible system testers, residents from different medical and surgical specialties should be recruited. Shortly, it is proposed to use IGlove with the residents of general surgery program, who are trained in performing endoscopic procedures using laparoscopic simulators [77].
During the test, rest periods that have been thought to simplify movement recognition process proved to be useful to simplify exercises segmentation, since during these periods the hand remains stationary on the table. However, placing a hand to a specific position to mark the beginning and end of a simulation activity can be impractical for particular types of surgical simulations. As one possible alternative, some authors propose to embed motion sensors inside the surgical instruments [37], in this way, periods of activity can be easily segmented, as they will coincide with those time intervals in which the instrument is being used.
The wrist cuff of the IGlove (i.e., the HUB), is undoubtedly an uncommon element to the traditional surgical practice. However, none of the participants stated that this instrument was uncomfortable. Part of the problem lies in its size because the electronic cards that were used for its implementation were Arduino compatible boards, which results in a rather bulky electronic device. One of the improvements proposed is the design of custom made cards, which will reduce the size of the HUB. However, the HUB cannot be avoided, since an embedded electronic system is required to extract the information acquired by each of the sensors in the glove. Other alternatives, such as providing wireless connectivity, would result in much more bulky sensors that could interfere with the normal hand movement since each sensor would need a wireless transmitter, a battery and a power management system. Another aspect to consider is the length of the cables connecting the sensors to the HUB. One of the alternatives tested was the use of longer cables to move the HUB to the arm. However, it was found that those long lines generated a severe degradation in high-speed digital signals, which forced to reduce their length to about 30 cm. Regardless, the HUB introduces a bias in all subjects, but as the variables are normalized by subject, it is fair to say, that each variable is compared intrasubject which implies that the bias is reduced in all subjects.
The experimental design can be considered useful for surgical simulation, as it was based on the Peg Transfer Board test, which is included in most of the laparoscopic simulators and is widely used for the evaluation of manual dexterity in endoscopic surgery [77]. It is also worth mentioning that such an exercise that involves handling tiny objects using small instruments in a reduced space resembles in many ways to some of the activities that are commonly performed by a surgeon during certain neurosurgical procedures. It was also found that sensors were sensitive enough to detect some kinds of events that may be unnoticeable to any visual evaluator. An example of this type of event is the momentary variation in the speed of rotation of the hand (which can be seen in the S ω signal like two peaks very close to each other). It was also noticed that some of the volunteers often did unnecessary movements when handling the tweezers (this can be seen on S ω signal as a long packet of small peaks). While these events were not used in the analysis reported in this paper, it is interesting to denote that these and other “hardly noticeable” events can be registered, and eventually be considered for further analysis.
As mentioned, all volunteers were filmed while they were doing the test. A careful review of video recording was performed to extract a set of metrics to feed an algorithm that clusters volunteers according to the level of manual dexterity that they have. In some sense, this human-driven evaluation can be considered an analogous of the assessment of an expert surgeon on simulated procedures. The Parameters that were shown in Table 1 and can be regarded as both subjective and objective in some ways. In the case of TSpan, TotalTask1, and TRatio, they were calculated by estimating the time duration of some activities, so, these could only present a small bias due to possible errors of appreciation by the person in charge of video review. Conversely, Errors are highly subjective, as the recognition of mistakes is subject to the personal judgment of the evaluator. The system was able to differentiate individuals according to their level of FMS, regardless of whether Errors parameter was taken into account or not; this is a remarkable fact indeed. This behavior can be attributed to the fact that people with higher manual dexterity made fewer mistakes (so their time parameters were better as they had no need to spend more time to correct their mistakes).
Results shown in Table 3 indicate that automatic measurements of the time duration of the third exercise are very similar to those obtained from the video review (column 2 in Table 1), the small differences can be attributed to errors of appreciation during the video review. The Metrics that were used to evaluate manual dexterity are stored in Table 3, Table 4, Table 5 and Table 6, in the columns that correspond to exercise 3. A trend can be seen among them; it is that volunteers with “lower numbers”, i.e., shorter duration of the exercises, fewer movements, and smaller displacement, tend to be the same, as expected, those participants with “lower numbers” were then classified in the group of high FMS. In the second and third columns in Table 4, it can be seen that all participants required almost the same amount of movement to complete exercises one and two. The analysis just mentioned confirms that the first two exercises do not measure any skill from participants (as these exercises were designed to test IGlove’s functionalities). By contrast, the third exercise can be considered a challenging activity where skilled volunteers can finish faster and using fewer movements than their counterparts.
Table 5 and Table 6 show spatial and angular displacements measured during the test. It can be noticed that displacements registered by all of the sensors on IGlove were very similar; this fact is especially noticeable for the case of exercises 1 and 3. Also on Table 7, the high values of the average of r and the small variability on them, reveals a strong correlation between movements registered by sensors for all the volunteers, it means that all sensors registered basically the same information during exercise 3. This fact can be attributed to the little difference in finger movements required to manipulate these small tweezers. However, on r 23 ( σ 23 ) , a little difference can be found, since r 23 values were slightly lower and at the same time σ 23 values were a bit higher than the values in the other columns. This variation may evidence the ability of the glove to detect the differences in movement resulting from small tool manipulation. Table 8 shows a similar behavior in the case of exercise 1, where most of the time all movements were very similar (since volunteers were instructed to put the hand on the table for a while just after performing each hand gesture). Only during exercise 2 sensors could register significant differences on movements. The previous results suggest that the information acquired from only one sensor is enough to assess of FMS in the third exercise (as the four sensors are reporting almost the same information).
The combination of the situation above discussed and the fact that it was found that the information from sensor fusion was not relevant for FMS of assessment, might suggest that the problem has been reduced to a point where the capabilities of IGlove are underused. However, such a drastic reduction is valid only for experimental designs where the movements of fingers relative to the hand are small. In those situations, where movements of fingers have a greater role, the information from sensors on the fingers must be taken into account (e.g., when utilizing laparoscopic instruments on the simulation of a minimally invasive surgery). Moreover, several authors had previously used only one sensor to assess manual dexterity [26,78].
As a remarkable fact, it can be said that based on the IGlove it has been possible to estimate some metrics that previously were only available in the most advanced simulators [8,27,79,80]:
  • The duration of each exercise is an important variable since it has been found that people with greater skills and experts can carry out simulation exercises faster. In the surgical field, the ability to perform procedures quickly, significantly reducing patient morbidity [81,82].
  • The number of movements performed is important, since it has been observed that people with greater ability can finish simulation exercises using fewer movements. In the surgical field, skilled surgeons can perform surgical procedures with a significant economy of movement (on real and simulated procedures), if compared to the number of movements (much higher) required by a resident [73,83,84].
  • Both spatial and angular displacement need to be taken into account, as skilled people tend to be able to perform simulation exercises with shorter paths and making fewer turns and rotations than novices [8,9,24]. In the surgical field, and specifically in neurosurgery, best paths are shorter. Surgeons that use optimal paths could reduce the incidence of some injuries caused by fatigue. Also from patient’s point-of-view, shorter paths can reduce injuries to brain tissue that might be due to the unnecessarily large amount of hand movement.
In Section 3.1 the results of the visual evaluation of dexterity were presented, volunteers were divided into groups according to the results given by an expert who reviewed the video records of the third exercise. Similarly, in Section 3.4 the results of the automatic evaluation of dexterity were presented, volunteers were divided into two groups according to the metrics obtained during the third exercise. It can be seen how in both cases, the group of trainees with “better numbers” was the same (subjects 1, 4, 5, 9, 12 and 13). At first glance, it seems that the IGlove could be just as capable as the expert in evaluating FMS. However, a conclusion like this one cannot be made only based on the results obtained with such a small sample of trainees. Larger longitudinal studies are required to corroborate or deny this possibility. Anyway, IGlove has not been conceived as a “replacement of professors”, but as a complementary tool for assessing residents’ FMS during their unsupervised practices.

5. Conclusions

Although gloves have been manufactured in different sizes using a thin and elastic fabric they are significantly different from latex or nitrile gloves conventionally used in neurosurgery. This difference in materials may adversely affect the performance of the trainees while training on the simulator. As a possible solution, it is proposed to put IMU sensors over traditional surgical gloves (e.g., using double sided tape), this will give a real tactile sensation to the trainees. Another element of IGlove that can significantly affect the development of the simulations is the HUB since this is not an element that a surgeon will find in an operating room. As previously shown the HUB is necessary for the IGlove to work, nevertheless, the current HUB is bulky so it can affect the performance of the trainee. Accordingly, a redesign of this system is necessary, focused on reducing the size of the HUB to the point where its interference may be considered neglectable.
Through the utilization of the tools described above and probably some new ones, innovative tests that incorporate aspects traditionally explored on Cognitive Neuroscience and Cognitive Psychology can be developed. Some of the mental processes studied by those disciplines include attention, visual memory, spatial perception, coordination, motricity, problem-solving, brain connectivity, among others. Those processes are intrinsically tied to the work of a neurosurgeon in his/her daily work. Most of the research in cognitive sciences is focused on studying children’s intellectual development or evaluating mentally-impaired individuals. However, the implementation of neuropsychological tests for healthy adults, and especially for highly skilled professionals (such as neurosurgeons) remains as an open research field.
An important limitation of this work is that IGlove’s metrics still need to be validated, to do that effectively, a comparison between IGlove’s results and metrics given by standard advanced simulators is necessary. In this regard, in an ongoing study, a group of trainees performs laparoscopic surgery training using a LAPSIM [19] and at the same time they wear an IGlove. The objective is to determine if IGlove metrics are good enough to evaluate the learning curve, compared with LAPSIM metrics which are taken as gold standards. However, the results obtained in this study were not included in this work, as the experiments have not finished yet, and are planned to be published in an upcoming paper.

Acknowledgments

This work has been supported partially by Vicerrectoría de Investigación of Universidad de Antioquia (CODI), Project “Sistema de Entrenamiento en Neurocirugía”, internal code MDC-10-1-6, and Fondo Nacional de Regalías de la República de Colombia through the project entitled “Fortalecimiento de la plataforma tecnológica para la especializada en el área de la salud y el desarrollo de la formación tecnología biomédica”, internal code 139C. The authors also want to thank W. Ramiréz, who provided the guidance for interpreting the results from a neurocognitive point of view.

Author Contributions

J.D. Lemos, A.M. Hérnandez, and G. Soto-Romero conceived the experiments; J.D. Lemos and G. Soto-Romero designed the experiments; J.D. Lemos performed the experiments; J.D. Lemos, A.M. Hérnandez, and G. Soto-Romero analyzed the data; J.D. Lemos wrote the paper with contributions from A.M. Hérnandez and G. Soto-Romero.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roitberg, B.; Banerjee, P.; Luciano, C.; Matulyauskas, M.; Rizzi, S.; Kania, P.; Gasco, J. Sensory and motor skill testing in neurosurgery applicants: A pilot study using a virtual reality haptic neurosurgical simulator. Neurosurgery 2013, 73, 116–121. [Google Scholar] [CrossRef] [PubMed]
  2. Hall, J.C.; Ellis, C.; Hamdorf, J. Surgeons and cognitive processes. Br. J. Surg. 2003, 90, 10–16. [Google Scholar] [CrossRef] [PubMed]
  3. Ackerman, P.L. A theory of adult intellectual development: Process, personality, interests, and knowledge. Intelligence 1996, 22, 227–257. [Google Scholar] [CrossRef]
  4. Loukas, C.; Georgiou, E. Multivariate autoregressive modeling of hand kinematics for laparoscopic skills assessment of surgical trainees. IEEE Trans. Biomed. Eng. 2011, 58, 3289–3297. [Google Scholar] [CrossRef] [PubMed]
  5. Van Hove, P.D.; Tuijthof, G.J.M.; Verdaasdonk, E.G.G.; Stassen, L.P.S.; Dankelman, J. Objective assessment of technical surgical skills. Br. J. Surg. 2010, 97, 972–987. [Google Scholar] [CrossRef] [PubMed]
  6. Ge, N. Surgical Expertise in Neurosurgery: Integrating Theory Into Practice. Neurosurgery 2013, 73, 30–38. [Google Scholar]
  7. Norcini, J.; Talati, J. Assessment, surgeon, and society. Int. J. Surg. 2009, 7, 313–317. [Google Scholar] [CrossRef] [PubMed]
  8. Alotaibi, F.E.; AlZhrani, G.A.; Sabbagh, A.J.; Azarnoush, H.; Winkler-Schwartz, A.; del Maestro, R.F. Neurosurgical Assessment of Metrics Including Judgment and Dexterity Using the Virtual Reality Simulator NeuroTouch (NAJD Metrics). Surg. Innov. 2015, 22. [Google Scholar] [CrossRef] [PubMed]
  9. Azarnoush, H.; Alzhrani, G.; Winkler-Schwartz, A.; Alotaibi, F.; Gelinas-Phaneuf, N.; Pazos, V.; Choudhury, N.; Fares, J.; DiRaddo, R.; del Maestro, R.F. Neurosurgical virtual reality simulation metrics to assess psychomotor skills during brain tumor resection. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 603–618. [Google Scholar] [CrossRef] [PubMed]
  10. Lemos, J.D.; Hernández, A.M.; Vallejo, A.F.; Estrada, D. System for Skills Training and Competency Assessment in Neurosurgery. In Proceedings of the 2013 Pan American Health Care Exchanges (Pahce 2013), Medellin, Colombia, 29 April–4 May 2013; p. 236. [Google Scholar]
  11. Tavakol, M.; Mohagheghi, M.A.; Dennick, R. Assessing the Skills of Surgical Residents Using Simulation. J. Surg. Educ. 2008, 65, 77–83. [Google Scholar] [CrossRef] [PubMed]
  12. Bann, S.; Darzi, A. Selection of individuals for training in surgery. Am. J. Surg. 2005, 190, 98–102. [Google Scholar] [CrossRef] [PubMed]
  13. Chin, K.J.; Tse, C.; Chan, V.; Tan, J.S.; Lupu, C.M.; Hayter, M. Hand motion analysis using the imperial college surgical assessment device: Validation of a novel and objective performance measure in ultrasound-guided peripheral nerve blockade. Reg. Anesth. Pain Med. 2011, 36, 213–219. [Google Scholar] [CrossRef] [PubMed]
  14. The Use of Simulation in Neurosurgical Education and Training. Available online: http://thejns.org/doi/pdf/10.3171/2014.5.JNS131766 (accessed on 29 April 2016).
  15. Reiley, C.E.; Lin, H.C.; Yuh, D.D.; Hager, G.D. Review of methods for objective surgical skill evaluation. Surg. Endosc. 2011, 25, 356–366. [Google Scholar] [CrossRef] [PubMed]
  16. Holloway, T.; Lorsch, Z.S.; Chary, M.A.; Sobotka, S.; Moore, M.M.; Costa, A.B.; Del, R.F.; Joshua, M. Operator experience determines performance in a simulated computer-based brain tumor resection task. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 1853–1862. [Google Scholar] [CrossRef] [PubMed]
  17. CAE Healthcare. Interventional Simulators: LapVR. Available online: http://caehealthcare.com/eng/interventional-simulators/lapvr (accessed on 23 May 2016).
  18. Introducing CAE Healthcare NeuroVR. Available online: http://caeneurovr.com/ (accessed on 18 May 2016).
  19. LapSim®: The Proven Training System. Available online: http://www.surgical-science.com/lapsim-the-proven-training-system/ (accessed on 11 May 2016).
  20. Malone, H.R.; Syed, O.N.; Downes, M.S.; D’Ambrosio, A.L.; Quest, D.O.; Kaiser, M.G. Simulation in neurosurgery: A review of computer-based simulation environments and their surgical applications. Neurosurgery 2010, 67, 1105–1116. [Google Scholar] [CrossRef] [PubMed]
  21. De Visser, H.; Watson, M.O.; Salvado, O.; Passenger, J.D. Progress in virtual reality simulators for surgical training and certification. Med. J. Austr. 2011, 194, 21–23. [Google Scholar]
  22. Hamza-lup, F.G.; Bogdan, C.M.; Popovici, D.M.; Costea, O.D. A Survey of Visuo-Haptic Simulation in Surgical Training. In Proceedings of the Third International Conference on Mobile, Hybrid, and on-Line Learning, Gosier, Guadeloupe, France, 23–28 February 2011; pp. 57–62. [Google Scholar]
  23. Alur, A.; Shrivastav, P.; Jumde, A. Haptic Technology: A Comprehensive Review of its Applications and Future Prospects. Int. J. Comput. Sci. Inf. Technol. 2014, 5, 6039–6043. [Google Scholar]
  24. Våpenstad, C.; Hofstad, E.F.; Bø, L.E.; Chmarra, M.K.; Kuhry, E.; Johnsen, G.; Mårvik, R.; Langø, T. Limitations of haptic feedback devices on construct validity of the LapSim® virtual reality simulator. Surg. Endosc. Other Interv. Tech. 2013, 27, 1386–1396. [Google Scholar] [CrossRef] [PubMed]
  25. Datta, V.; Mackay, S.; Mandalia, M.; Darzi, A. The use of electromagnetic motion tracking analysis to objectively measure open surgical skill in the laboratory-based model. J. Am. Coll. Surg. 2001, 193, 479–485. [Google Scholar] [CrossRef]
  26. Datta, V.; Chang, A.; Mackay, S.; Darzi, A. The relationship between motion analysis and surgical technical assessments. Am. J. Surg. 2002, 184, 70–73. [Google Scholar] [CrossRef]
  27. Rolls, A.E.; Riga, C.V.; Bicknell, C.D.; Stoyanov, D.V.; Shah, C.V.; van Herzeele, I.; Hamady, M.; Cheshire, N.J. A pilot study of video-motion analysis in endovascular surgery: Development of real-time discriminatory skill metrics. Eur. J. Vasc. Endovasc. Surg. 2013, 45, 509–515. [Google Scholar] [CrossRef] [PubMed]
  28. Hillebrand, G.; Bauer, M.; Achatz, K. Inverse kinematic infrared optical finger tracking. In Proceedings of the 9th International Conference on Humans and Computers (HC 2006), Aizu, Japan, 6–9 September 2006; pp. 1–6. [Google Scholar]
  29. Du, G.; Zhang, P.; Liu, X. Markerless Human-Manipulator Interface Using Leap Motion with Interval Kalman Filter and Improved Particle Filter. IEEE Trans. Ind. Inform. 2016, 12, 694–704. [Google Scholar] [CrossRef]
  30. Chen, W.; Xiong, C.; Huang, X.; Sun, R.; Xiong, Y. Kinematic analysis and dexterity evaluation of upper extremity in activities of daily living. Gait Posture 2010, 32, 475–481. [Google Scholar] [CrossRef] [PubMed]
  31. Metcalf, C.D.; Notley, S.V.; Chappell, P.H.; Burridge, J.H.; Yule, V.T. Validation and application of a computational model for wrist and hand movements using surface markers. Biomed. Eng. IEEE Trans. 2008, 55, 1199–1210. [Google Scholar] [CrossRef] [PubMed]
  32. Zappella, L.; Béjar, B.; Hager, G.; Vidal, R. Surgical gesture classification from video and kinematic data. Med. Image Anal. 2013, 17, 732–745. [Google Scholar] [CrossRef] [PubMed]
  33. Gray, R.J.; Kahol, K.; Islam, G.; Smith, M.; Chapital, A.; Ferrara, J. High-fidelity, low-cost, automated method to assess laparoscopic skills objectively. J. Surg. Educ. 2012, 69, 335–339. [Google Scholar] [CrossRef] [PubMed]
  34. Uemura, M.; Tomikawa, M.; Kumashiro, R.; Miao, T.; Souzaki, R.; Ieiri, S.; Ohuchida, K.; Lefor, A.T.; Hashizume, M. Analysis of hand motion differentiates expert and novice surgeons. J. Surg. Res. 2014, 188, 8–13. [Google Scholar] [CrossRef] [PubMed]
  35. Harada, K.; Morita, A.; Minakawa, Y.; Baek, Y.M.; Sora, S.; Sugita, N.; Kimura, T.; Tanikawa, R.; Ishikawa, T.; Mitsuishi, M. Assessing Microneurosurgical Skill with Medico-Engineering Technology. World Neurosurg. 2015, 84, 964–971. [Google Scholar] [CrossRef] [PubMed]
  36. Hammond, F.L.; Smith, M.J.; Wood, R.J. Printing Strain Gauges on Surgical Instruments for Force Measurement. J. Med. Device 2014, 8, 30935. [Google Scholar] [CrossRef]
  37. Hammond, F.L.; Talbot, S.G.; Wood, R.J.; Howe, R.D. Measurement System for the Characterization of Micro- Manipulation Motion and Force. J. Med. Device 2013, 7, 0309401–0309402. [Google Scholar] [CrossRef]
  38. Overby, D.W.; Watson, R.A. Hand motion patterns of Fundamentals of Laparoscopic Surgery certified and noncertified surgeons. Am. J. Surg. 2014, 207, 226–230. [Google Scholar] [CrossRef] [PubMed]
  39. Watson, R.A. Quantification of Surgical Technique Using an Inertial Measurement Unit. Simul. Healthc. 2013, 8, 162–165. [Google Scholar] [CrossRef] [PubMed]
  40. Ohu, I.; Cho, S.; Zihni, A.; Cavallo, J.A.; Awad, M.M. Analysis of surgical motions in minimally invasive surgery using complexity theory. Int. J. Biomed. Eng. Technol. 2015, 17, 24–41. [Google Scholar] [CrossRef]
  41. What Gestures Does the Myo Armband Recognize? Available online: https://support.getmyo.com/hc/en-us/articles/202647853-What-gestures-does-the-Myo-armband-recognize- (accessed on 18 May 2016).
  42. Camacho Navarro, J.; Leon-Vargas, F.; Barrero Perez, J. EMG-Based System for Basic Hand Movement Recognition. Dyna 2012, 79, 41–49. [Google Scholar]
  43. Virtual Motion Labs. Available online: http://www.virtualmotionlabs.com/ (accessed on 12 May 2016).
  44. CyberGlove Systems. Available online: http://www.cyberglovesystems.com/ (accessed on 20 May 2016).
  45. Dipietro, L.; Sabatini, A.M.; Dario, P. A survey of glove-based systems and their applications. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2008, 38, 461–482. [Google Scholar] [CrossRef]
  46. Giuffrida, J.P.; Riley, D.E.; Maddux, B.N.; Heldmann, D.A. Clinically deployable kinesia technology for automated tremor assessment. Mov. Disord. 2009, 24, 723–730. [Google Scholar] [CrossRef] [PubMed]
  47. Dai, H.; Otten, B.; Mehrkens, J.H.; D’Angelo, L.T.; Lueth, T.C. A novel glove monitoring system used to quantify neurological symptoms during deep-brain stimulation surgery. IEEE Sens. J. 2013, 13, 3193–3202. [Google Scholar] [CrossRef]
  48. Friedman, N.; Chan, V.; Reinkensmeyer, A.N.; Beroukhim, A.; Zambrano, G.J.; Bachman, M.; Reinkensmeyer, D.J. Retraining and assessing hand movement after stroke using the MusicGlove: Comparison with conventional hand therapy and isometric grip training. J. Neuroeng. Rehabil. 2014, 11, 76. [Google Scholar] [CrossRef] [PubMed]
  49. Li, K.; Chen, I.M.; Yeo, S.H.; Lim, C.K. Development of finger-motion capturing device based on optical linear encoder. J. Rehabil. Res. Dev. 2011, 48, 69–82. [Google Scholar] [CrossRef] [PubMed]
  50. Nishiyama, M.; Watanabe, K. Wearable sensing glove with embedded hetero-core fiber-optic nerves for unconstrained hand motion capture. IEEE Trans. Instrum. Meas. 2009, 58, 3995–4000. [Google Scholar] [CrossRef]
  51. Tanyawiwat, N.; Thiemjarus, S. Design of an Assistive Communication Glove Using Combined Sensory Channels. In Proceedings of the 2012 Ninth International Conference on Wearable and Implantable Body Sensor Networks (BSN), London, UK, 9–12 May 2012; pp. 34–39. [Google Scholar]
  52. Gentner, R.; Classen, J. Development and evaluation of a low-cost sensor glove for assessment of human finger movements in neurophysiological settings. J. Neurosci. Methods 2009, 178, 138–147. [Google Scholar] [CrossRef] [PubMed]
  53. Fahn, C.S.; Sun, H. Development of a Data Glove With Reducing Sensors Based on Magnetic Induction. IEEE Trans. Ind. Electron. 2005, 52, 585–594. [Google Scholar] [CrossRef]
  54. Kindratenko, V.V. A survey of electromagnetic position tracker calibration techniques. Virtual Real. 2000, 5, 169–182. [Google Scholar] [CrossRef]
  55. Sessa, S.; Zecca, M.; Lin, Z.; Sasaki, T.; Suzuki, T.; Itoh, K.; Iseki, H.; Takanishi, A. Objective skill analysis and assessment of neurosurgery by using the waseda bioinstrumentation system WB-3. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, St. Louis Riverfront, MO, USA, 10–15 October 2009; pp. 4086–4091. [Google Scholar]
  56. Saber-Sheikh, K.; Bryant, E.C.; Glazzard, C.; Hamel, A.; Lee, R.Y.W. Feasibility of using inertial sensors to assess human movement. Man. Ther. 2010, 15, 122–125. [Google Scholar] [CrossRef] [PubMed]
  57. Xsense Corporation. Human Motion Measurement. Available online: https://www.xsens.com/functions/human-motion-measurement/ (accessed on 11 May 2016).
  58. Lemos, J.D.; Hernández, A.M.; Soto-Romero, G.; Valade, A. Instrumented glove for in-hand movement tracking in neurosurgical simulation. In Proceedings of the 2014 21st IEEE International Conference on Electronics, Circuits and Systems (ICECS), Marseille, France, 7–10 December 2014. [Google Scholar]
  59. Lemos, J.D.; Hernández, A.M.; Soto-Romero, G.; Valade, A. Instrumented Glove for Skills Assessment in Neurosurgical Simulation. In Proceedings of the Biomedical Circuits and Systems Conference (BioCAS), Lausanne, Switzerland, 22–24 October 2014; pp. 308–311. [Google Scholar]
  60. C. Corp. CY8C4245AXI-483. Available online: http://www.cypress.com/part/cy8c4245axi-483 (accessed on 27 May 2016).
  61. I. Corp. MPU-9250 Nine-Axis (Gyro + Accelerometer + Compass) MEMS MotionTrackingTM Device. Available online: https://www.invensense.com/products/motion-tracking/9-axis/mpu-9250/ (accessed on 11 May 2016).
  62. O. Corp. Polarized Nano Connectors. Available online: http://www.omnetics.com/polarized-nano/cots/ (accessed on 11 May 2016).
  63. RehabMart: Manipulation/Dexterity Tests, Functional Dexterity Text, Purdue Pegboard Test, Fine Motor Skills. Available online: http://www.rehabmart.com/category/manipulation/dexterity_tests.htm (accessed on 28 May 2016).
  64. Strauss, E.; Sherman, E.; Spreen, O. A Compendium of Neuropsychological Tests: Administration, Norms, and Commentary. In A Compendium of Neuropsychological Tests: Administration, Norms, and Commentary, 3rd ed.; Oxford University Press: New York, NY, USA, 2006; Chapers 12–14. [Google Scholar]
  65. Sadideen, H.; Alvand, A.; Saadeddin, M.; Kneebone, R. Surgical experts: Born or made? Int. J. Surg. 2013, 11, 773–778. [Google Scholar] [CrossRef] [PubMed]
  66. Van Herzeele, I.; O’Donoghue, K.G.L.; Aggarwal, R.; Vermassen, F.; Darzi, A.; Cheshire, N.J.W. Visuospatial and psychomotor aptitude predicts endovascular performance of inexperienced individuals on a virtual reality simulator. J. Vasc. Surg. 2010, 51, 1035–1042. [Google Scholar] [CrossRef] [PubMed]
  67. Jardine, D.; Hoagland, B.; Perez, A.; Gessler, E. Evaluation of Surgical Dexterity During the Interview Day: Another Factor for Consideration. J Grad. Med. Educ. 2015, 7, 234–237. [Google Scholar] [CrossRef] [PubMed]
  68. Wanzel, K.R.; Hamstra, S.J.; Caminiti, M.F.; Anastakis, D.J.; Grober, E.D.; Reznick, R.K. Visual-spatial ability correlates with efficiency of hand motion and successful surgical performance. Surgery 2003, 134, 750–757. [Google Scholar] [CrossRef]
  69. C. Simulab. Peg Transfer Board (with Triangles). Available online: https://www.simulab.com/products/peg-transfer-board-triangles (accessed on 25 May 2016).
  70. Arikatla, V.S.; Sankaranarayanan, G.; Ahn, W.; Chellali, A.; Cao, C.G.L.; De, S. Development and validation of VBLaST-PT©: A virtual peg transfer simulator. Stud. Health Technol. Inform. 2013, 184, 24–30. [Google Scholar] [PubMed]
  71. Madgwick, S.O.H.; Harrison, A.J.L.; Vaidyanathan, R. Estimation of IMU and MARG orientation using a gradient descent algorithm. In Proceedings of the IEEE International Conference on Rehabilitation Robotics, Zurich, Switzerland, 29 June–1 July 2011; pp. 978–984. [Google Scholar]
  72. Olivares, A.; Ramirez, J.; Gorriz, J.M.; Olivares, G.; Damas, M. Detection of (In)activity Periods in Human Body Motion Using Inertial Sensors: A Comparative Study. Sensors 2012, 12, 5791–5814. [Google Scholar] [CrossRef] [PubMed]
  73. Stančin, S.; Tomažič, S. Early improper motion detection in golf swings using wearable motion sensors: The first approach. Sensors 2013, 13, 7505–7521. [Google Scholar] [CrossRef] [PubMed]
  74. Moorthy, K.; Munz, Y.; Sarker, S.K.; Darzi, A. Clinical review Objective assessment of technical skills in surgery. BMJ 2003, 327, 1032–1037. [Google Scholar] [CrossRef] [PubMed]
  75. Ozyagcilar, T. Calibrating an eCompass in the Presence of Hard and Soft-Iron Interference; Freescale Semiconductor Ltd.: Austin, TX, USA, 2012. [Google Scholar]
  76. Loyo, M. Neurosurgery in Mexico and Latin-America. Int. Neurosci. J. 2015, 1, 45–46. [Google Scholar] [CrossRef]
  77. Poolton, J.M.; Wilson, M.R.; Malhotra, N.; Ngo, K.; Masters, R.S.W. A comparison of evaluation, time pressure, and multitasking as stressors of psychomotor operative performance. Surgery 2011, 149, 776–782. [Google Scholar] [CrossRef] [PubMed]
  78. Murphy, T.E. Towards Objective Surgical Skill Evaluation with Hidden Markov Model-Based Motion Recognition; Johns Hopkins University: Baltimore, MD, USA, 2004. [Google Scholar]
  79. Gallagher, A.G. Metric-based simulation training to proficiency in medical education: What it is and how to do it. Ulster Med. J. 2012, 81, 107–113. [Google Scholar] [PubMed]
  80. Sabbagh, A.J.; Winkler-Schwartz, A.; del Maestro, R.F. Assessing Bimanual Performance in Brain Tumor Resection with NeuroTouch, a Virtual Reality Simulator. Oper. Neurosurg. 2015, 11, 89–98. [Google Scholar]
  81. Lin, Z.; Zecca, M.; Sessa, S.; Sasaki, T.; Suzuki, T.; Itoh, K.; Iseki, H.; Takanishi, A. Objective skill analysis and assessment in neurosurgery by using an ultra-miniaturized inertial measurement unit WB-3—Pilot tests. In Proceedings of the Annual International Conference of Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 2–6 September 2009; pp. 2320–2323. [Google Scholar]
  82. Deml, B.; Ortmaier, T.; Seibold, U. The touch and feel in minimally invasive surgery. IREE Int. Work. Haptic Audio Vis. Environ. Their Appl. 2005, 2005, 33–38. [Google Scholar]
  83. Reznick, R.; Regehr, G.; MacRae, H.; Martin, J.; McCulloch, W. Testing technical skill via an innovative ‘bench station’ examination. Am. J. Surg. 1997, 173, 226–230. [Google Scholar] [CrossRef]
  84. Datta, V.; Bann, S.; Mandalia, M.; Darzi, A. The surgical efficiency score: A feasible, reliable, and valid method of skills assessment. Am. J. Surg. 2006, 192, 372–378. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Hardware of IGlove system: (a) Block diagram of main parts of Hub and Glove and their interconnection. Microcontroller and sensors communicate in a single-master, multiple-slave configuration using SPI on a multidrop bus. IMU sensors were configured to sample internal signals from accelerometers and gyroscopes at 8 KHz, and the user-programmable digital low-pass filters were setting up to a cut-off frequency of 125 Hz, using the preconfigured low-pass filters of the sensor. Measurement data is sent via a high-speed USB link to the Daubara NS Trainer; sensors are sampled 100 times per second; (b) Picture of the system showing the connection between Glove and Hub. In this case, sensors were worn on thumb, index, middle, ring fingers, and in the back of the hand. A communication system that transfers data and power between Hub and each sensor has been implemented using a set of thin and flexible cables. A zoom of the thumb is presented in the lower right corner; it can be observed how power and data cables enter to one sensor compartment.
Figure 1. Hardware of IGlove system: (a) Block diagram of main parts of Hub and Glove and their interconnection. Microcontroller and sensors communicate in a single-master, multiple-slave configuration using SPI on a multidrop bus. IMU sensors were configured to sample internal signals from accelerometers and gyroscopes at 8 KHz, and the user-programmable digital low-pass filters were setting up to a cut-off frequency of 125 Hz, using the preconfigured low-pass filters of the sensor. Measurement data is sent via a high-speed USB link to the Daubara NS Trainer; sensors are sampled 100 times per second; (b) Picture of the system showing the connection between Glove and Hub. In this case, sensors were worn on thumb, index, middle, ring fingers, and in the back of the hand. A communication system that transfers data and power between Hub and each sensor has been implemented using a set of thin and flexible cables. A zoom of the thumb is presented in the lower right corner; it can be observed how power and data cables enter to one sensor compartment.
Sensors 17 00988 g001
Figure 2. Sensor compartments and a motion sensor: (a) Distribution of sensor compartments on IGlove. Red arrows point to compartments used to accommodate the four sensors employed in this study; Sensor 1 is in the back of the hand, Sensor 2 on the thumb, Sensor 3 on the index, and Sensor 4 on the middle finger. 16 gray marks on glove indicate available places to put motion sensors (The glove has 16 compartments to put sensors, three for each finger on each phalange, and one in the back of the hand); (b) A sensor beside a 6 mm pencil to illustrate size. This sensor finely fits on any of the compartments on the Glove.
Figure 2. Sensor compartments and a motion sensor: (a) Distribution of sensor compartments on IGlove. Red arrows point to compartments used to accommodate the four sensors employed in this study; Sensor 1 is in the back of the hand, Sensor 2 on the thumb, Sensor 3 on the index, and Sensor 4 on the middle finger. 16 gray marks on glove indicate available places to put motion sensors (The glove has 16 compartments to put sensors, three for each finger on each phalange, and one in the back of the hand); (b) A sensor beside a 6 mm pencil to illustrate size. This sensor finely fits on any of the compartments on the Glove.
Sensors 17 00988 g002
Figure 3. Hand gestures (1 to 9) used in exercises one and two. Those gestures were designed to test the capability of IGlove system for recognizing finger movements, using sensors on the thumb, index, and middle fingers.
Figure 3. Hand gestures (1 to 9) used in exercises one and two. Those gestures were designed to test the capability of IGlove system for recognizing finger movements, using sensors on the thumb, index, and middle fingers.
Sensors 17 00988 g003
Figure 4. The third exercise of IGlove test: (a) Shows the PegBoard, tweezers holding a plastic bushing (PB) and a portion of PBs deposits on the far right of the image. The PegBoard features many pegs, but subjects only need to put six blue PBs on blue pins and one red PB on each peak of the star; (b) A picture of one of the volunteers while running the third exercise (she is just placing a blue PB on a blue peg).
Figure 4. The third exercise of IGlove test: (a) Shows the PegBoard, tweezers holding a plastic bushing (PB) and a portion of PBs deposits on the far right of the image. The PegBoard features many pegs, but subjects only need to put six blue PBs on blue pins and one red PB on each peak of the star; (b) A picture of one of the volunteers while running the third exercise (she is just placing a blue PB on a blue peg).
Sensors 17 00988 g004
Figure 5. Scheme of signal processing and information analysis to obtain manual dexterity metrics on each sensor. Currently, four metrics have been achieved from IGlove’s raw signals: Duration of exercises (Duration), the number of movements during each exercise (Movements), spatial displacement (Disp), and angular displacement ( D i s p θ ). The absolute value of the Hilbert transform of a gyroscope signal is filtered to obtain a smoothened signal that is higher than 0 when the hand moves. A comparator decides when the level and duration of a perturbation in the signal were enough to mark it as an exercise. The metric Duration of an exercise is the difference between the end and the beginning of a significant perturbation. Variable Movements is calculated from the magnitude of angular speed S ω (computed from all gyroscopes using Equation (1)). It was observed that peaks on S ω indicate when a movement was performed. To estimate the number of moves, S ω is smoothened, but this time, a peak detector identifies the spikes in the signal that were large enough and have been caused by a valid hand movement. The area under S ω was calculated using numerical integration to estimate D i s p θ . Spatial kinematic parameters were estimated from accelerometers signal S a (calculated using Equation (3)), Runge–Kutta method was used to obtain an approximate time series for the magnitude of velocity V a . Finally, spatial displacement is calculated as the area under the curve of | V a ( t ) | . When performing numerical integration, Duration and Movements parameters were used to establish integration limits and estimate initial conditions.
Figure 5. Scheme of signal processing and information analysis to obtain manual dexterity metrics on each sensor. Currently, four metrics have been achieved from IGlove’s raw signals: Duration of exercises (Duration), the number of movements during each exercise (Movements), spatial displacement (Disp), and angular displacement ( D i s p θ ). The absolute value of the Hilbert transform of a gyroscope signal is filtered to obtain a smoothened signal that is higher than 0 when the hand moves. A comparator decides when the level and duration of a perturbation in the signal were enough to mark it as an exercise. The metric Duration of an exercise is the difference between the end and the beginning of a significant perturbation. Variable Movements is calculated from the magnitude of angular speed S ω (computed from all gyroscopes using Equation (1)). It was observed that peaks on S ω indicate when a movement was performed. To estimate the number of moves, S ω is smoothened, but this time, a peak detector identifies the spikes in the signal that were large enough and have been caused by a valid hand movement. The area under S ω was calculated using numerical integration to estimate D i s p θ . Spatial kinematic parameters were estimated from accelerometers signal S a (calculated using Equation (3)), Runge–Kutta method was used to obtain an approximate time series for the magnitude of velocity V a . Finally, spatial displacement is calculated as the area under the curve of | V a ( t ) | . When performing numerical integration, Duration and Movements parameters were used to establish integration limits and estimate initial conditions.
Sensors 17 00988 g005
Figure 6. Sections of S ω acquired from sensor 1 (in the back of the hand), (a) Representation of peaks on S ω signal on third exercise: The first peak corresponds to the time where a PB is collected from deposit, the second peak, corresponding to the instant at which the PB is placed in the corresponding peg and, the third peak corresponds to the action of taking the next PB. Task1 occurred between the first and second peak, and Task2 occurred between the second and third peak; (b) Some particular cases of peaks that can be found in the S ω signal. Arrow 1: shows a peak too small to be considered a voluntary movement (to be discarded), Arrows 2: shows sets of nearby peaks that represent single movements, and arrow 3, shows a much smaller peak, but not small enough to be eliminated (it represents a hand movement).
Figure 6. Sections of S ω acquired from sensor 1 (in the back of the hand), (a) Representation of peaks on S ω signal on third exercise: The first peak corresponds to the time where a PB is collected from deposit, the second peak, corresponding to the instant at which the PB is placed in the corresponding peg and, the third peak corresponds to the action of taking the next PB. Task1 occurred between the first and second peak, and Task2 occurred between the second and third peak; (b) Some particular cases of peaks that can be found in the S ω signal. Arrow 1: shows a peak too small to be considered a voluntary movement (to be discarded), Arrows 2: shows sets of nearby peaks that represent single movements, and arrow 3, shows a much smaller peak, but not small enough to be eliminated (it represents a hand movement).
Sensors 17 00988 g006
Figure 7. Clustering results according to the variables evaluated from video records of volunteers while executing exercise 3. (a) Clustering classification using variables: TSpan, TotalTask1, and TRatio. Errors were not utilized in this classification process, to evaluate if it introduces subjective bias (Estimating the number of errors is dependent on the judgment of the evaluator); (b) Hierarchical classification using variables: variables TSpan, TotalTask1, TRatio, and Errors. Three groups can be clearly differentiated, blue (volunteers 1, 4, 5, 9, 12 and 13), red (volunteers 7, 8, 11 and 14) and green (volunteers 2, 3, 6 and 10); volunteer 10 can be considered either green or red.
Figure 7. Clustering results according to the variables evaluated from video records of volunteers while executing exercise 3. (a) Clustering classification using variables: TSpan, TotalTask1, and TRatio. Errors were not utilized in this classification process, to evaluate if it introduces subjective bias (Estimating the number of errors is dependent on the judgment of the evaluator); (b) Hierarchical classification using variables: variables TSpan, TotalTask1, TRatio, and Errors. Three groups can be clearly differentiated, blue (volunteers 1, 4, 5, 9, 12 and 13), red (volunteers 7, 8, 11 and 14) and green (volunteers 2, 3, 6 and 10); volunteer 10 can be considered either green or red.
Sensors 17 00988 g007
Figure 8. Segmentation of exercises. (a) Raw signal from one Z-axis gyroscope from Sensor 1 on the back of the hand ( S G 1 z ); (b) the envelope of the previous signal calculated using the transform Gilbert (green). The exercise detection signal (in blue), which is basically the output of a comparator, with values of 100% when | S G 1 z ^ | > threshold (exercise is detected) and 0% when | S G 1 z ^ | < threshold (rest periods).
Figure 8. Segmentation of exercises. (a) Raw signal from one Z-axis gyroscope from Sensor 1 on the back of the hand ( S G 1 z ); (b) the envelope of the previous signal calculated using the transform Gilbert (green). The exercise detection signal (in blue), which is basically the output of a comparator, with values of 100% when | S G 1 z ^ | > threshold (exercise is detected) and 0% when | S G 1 z ^ | < threshold (rest periods).
Sensors 17 00988 g008
Figure 9. Peak detection results in two trainees during the third exercise. In both cases, the raw signal of magnitude of angular velocity S ω is painted in magenta. The S ω signal after being “smoothed” | S ω ^ | is painted blue. Detected peaks are marked with green triangles. (a) Subject 12, required only 25 movements to finish the third exercise. It can be seen how the good level of manual dexterity is also reflected in the in the smoothness of the signal acquired by the sensors in the glove; By contrast, subject 2 (b) made many more moves during the exercise; this was because he made several mistakes and then had to correct them. The significant differences in movement amplitude may indicate poor motor control.
Figure 9. Peak detection results in two trainees during the third exercise. In both cases, the raw signal of magnitude of angular velocity S ω is painted in magenta. The S ω signal after being “smoothed” | S ω ^ | is painted blue. Detected peaks are marked with green triangles. (a) Subject 12, required only 25 movements to finish the third exercise. It can be seen how the good level of manual dexterity is also reflected in the in the smoothness of the signal acquired by the sensors in the glove; By contrast, subject 2 (b) made many more moves during the exercise; this was because he made several mistakes and then had to correct them. The significant differences in movement amplitude may indicate poor motor control.
Sensors 17 00988 g009
Figure 10. Graph of kinematic spatial parameters for volunteer 12 while playing exercise 3. (a) Black signal S a is the magnitude of the acceleration, the absolute value of velocity | V a ( t ) | is drawn in green, and the area under the curve (in clear yellow) is D i s p ; (b) corresponds to the peak detection of S ω , each pair of peaks were used to set the appropriate time interval for numerical calculations.
Figure 10. Graph of kinematic spatial parameters for volunteer 12 while playing exercise 3. (a) Black signal S a is the magnitude of the acceleration, the absolute value of velocity | V a ( t ) | is drawn in green, and the area under the curve (in clear yellow) is D i s p ; (b) corresponds to the peak detection of S ω , each pair of peaks were used to set the appropriate time interval for numerical calculations.
Sensors 17 00988 g010
Figure 11. Comparison of techniques to calculate angular speed. (a) Angular velocity calculated from absolute orientation angles ϕ , θ and ψ , obtained using a sensor fusion algorithm; (b) Angular speed calculated from raw gyroscope data.
Figure 11. Comparison of techniques to calculate angular speed. (a) Angular velocity calculated from absolute orientation angles ϕ , θ and ψ , obtained using a sensor fusion algorithm; (b) Angular speed calculated from raw gyroscope data.
Sensors 17 00988 g011
Figure 12. Dendrograms from cluster analysis of volunteers according to variables from data processing of IGlove data. (a) Clustering using Time of Exercise, Peaks quantity, spatial displacement and angular displacement; (b) Clustering using Peaks quantity, spatial displacement, and angular displacement.
Figure 12. Dendrograms from cluster analysis of volunteers according to variables from data processing of IGlove data. (a) Clustering using Time of Exercise, Peaks quantity, spatial displacement and angular displacement; (b) Clustering using Peaks quantity, spatial displacement, and angular displacement.
Sensors 17 00988 g012
Table 1. Results for each of the volunteers obtained from the review of videos recorded during the third exercise. The Classification column shows resulting hierarchical classification from dendrograms shown in Figure 7.
Table 1. Results for each of the volunteers obtained from the review of videos recorded during the third exercise. The Classification column shows resulting hierarchical classification from dendrograms shown in Figure 7.
VolunteerTSpanTotalTask1TRatioErrors
142.224.91.692
277.733.72.3110
360.733.41.824
44827.11.774
548.421.72.233
661.230.91.986
765.122.32.919
8107.1254.2817
952.627.41.923
1082.328.42.912
1174.4252.989
1234.418.11.90
1346.922.92.056
1465.618.43.579
Table 2. Comparison of the level of dexterity found in each group. Each column represents the average and standard deviations for each variable in each group. The blue group has high dexterity, as they, in general, take shorter times, lower TRatio values and made fewer errors. Conversely, the red group has lower dexterity, as they in general, take longer times, higher TRatio values and made many errors. The green group has average dexterity.
Table 2. Comparison of the level of dexterity found in each group. Each column represents the average and standard deviations for each variable in each group. The blue group has high dexterity, as they, in general, take shorter times, lower TRatio values and made fewer errors. Conversely, the red group has lower dexterity, as they in general, take longer times, higher TRatio values and made many errors. The green group has average dexterity.
Group T S p a n ¯   ( σ T S p a n ) T o t a l T a s k 1 ¯   ( σ T o t a l T a s k 1 ) T R a t i o ¯   ( σ T R a t i o ) E r r o r s ¯   ( σ E r r o r s )
Blue45.42 (6.34)23.68 (3.54)1.93 (0.19)3 (2.00)
Green70.48 (11.16)31.60 (2.48)2.25 (0.48)8 (3.65)
Red78.05 (19.83)22.68 (3.12)3.44 (0.64)11 (4.00)
Table 3. Automatic measurements of the duration of the exercises, time, is represented in seconds.
Table 3. Automatic measurements of the duration of the exercises, time, is represented in seconds.
VolunteerExercise 1Exercise 2Exercise 3
144.021.643.1
233.816.378.6
350.929.560.7
423.610.847.7
540.919.549.4
637.619.260
742.329.965.1
825.415.3106.3
946.124.551.8
1036.717.877.9
1143.120.074.4
1228.417.334.9
1323.513.548.3
1429.515.866.6
Table 4. Automatic detection of movements. Each column shows the number of movements carried out by a volunteer to complete an exercise. Exercises 1 and 2 did not represent any challenge for any of the participants as exercises were carried out with almost the same quantity of movements. On the contrary, the results of the third exercise were very heterogeneous; some participants required few movements to do exercise while others require many more movements. Extra-movements (more than 25) can be attributed to errors or unnecessary moves during the exercise execution.
Table 4. Automatic detection of movements. Each column shows the number of movements carried out by a volunteer to complete an exercise. Exercises 1 and 2 did not represent any challenge for any of the participants as exercises were carried out with almost the same quantity of movements. On the contrary, the results of the third exercise were very heterogeneous; some participants required few movements to do exercise while others require many more movements. Extra-movements (more than 25) can be attributed to errors or unnecessary moves during the exercise execution.
VolunteerExercise 1Exercise 2Exercise 3
1191132
2191151
3201035
4191033
5191131
6191147
7191056
8191071
9191032
10211060
11201056
12191025
13201037
14191051
Table 5. Spatial displacement (Disp) measured by all sensors on each of the exercises. E1 means Exercise 1; E2 Exercise 2; and E3 Exercise 3. In all cases, displacements are expressed in meters.
Table 5. Spatial displacement (Disp) measured by all sensors on each of the exercises. E1 means Exercise 1; E2 Exercise 2; and E3 Exercise 3. In all cases, displacements are expressed in meters.
Sensor 1Sensor 2Sensor 3Sensor 4
VolE1E2E3E1E2E3E1E2E3E1E2E3
13.632.721.943.913.032.053.672.901.983.842.992.04
25.171.345.25.381.595.315.121.495.245.281.605.31
34.761.803.165.052.023.294.821.903.214.991.983.26
44.411.712.694.652.002.784.381.902.704.581.992.77
53.450.862.123.641.132.233.420.982.153.581.102.21
65.162.802.275.343.112.385.112.992.315.313.102.37
73.921.302.414.221.612.523.941.462.444.131.572.49
81.913.124.042.123.414.041.883.273.962.033.384.03
93.493.692.353.704.022.403.423.892.343.583.992.40
102.911.013.253.181.243.332.921.133.273.111.203.32
113.841.873.394.092.103.513.842.003.434.032.103.48
123.411.572.023.701.892.073.451.742.003.601.832.06
136.272.162.326.512.402.416.272.272.346.422.392.41
142.430.972.752.661.192.822.401.092.752.571.192.82
Table 6. Angular displacement ( D i s p θ ) measured by all sensors on each of the exercises. E1 means Exercise 1; E2 Exercise 2; and E3 Exercise 3. In all cases, angular displacements are expressed in degrees.
Table 6. Angular displacement ( D i s p θ ) measured by all sensors on each of the exercises. E1 means Exercise 1; E2 Exercise 2; and E3 Exercise 3. In all cases, angular displacements are expressed in degrees.
Sensor 1Sensor 2Sensor 3Sensor 4
VolE1E2E3E1E2E3E1E2E3E1E2E3
117885151387223476915651803523141521017271568
220495542578256287928542043581258423188112787
314135402080179781522721429558205216017722237
41557731143219251104159815617351442179810191545
514244569241715708101214024789161649666991
615855321244199679513681575534124018127311340
715304991776189778218951541523176417347201936
8160166618691980100020321584671185417889352031
915864481184197670112971577464119818356471299
1012085151415153279615521214530142614307451559
1114925551499191786217141519577152017668091671
121795574565221589261218046005572085832602
131734841110322391288118717548701099201512201188
1414453641401178954315471454369141816555151519
Table 7. r calculated between sensor 1 and sensors 2, 3 and 4 during exercise 3. Parameter r 12 is the average of r calculated between the signals S ω 1 and S ω 2 over all the movements that each volunteer performed to complete the exercise 3; ( σ 12 ) is the standard deviation of these calculations. r 13 ( σ 13 ) , gives the same information as previous column, but correlation was performed between sensors 1 and 3. r 14 ( σ 14 ) , In the same way as the previous two, brings the information regarding correlation between sensor 1 and 4. Finally, r 23 ( σ 23 ) shows behavior of the combination thumb-index (for exercise 3, it brings information regarding tweezers manipulation).
Table 7. r calculated between sensor 1 and sensors 2, 3 and 4 during exercise 3. Parameter r 12 is the average of r calculated between the signals S ω 1 and S ω 2 over all the movements that each volunteer performed to complete the exercise 3; ( σ 12 ) is the standard deviation of these calculations. r 13 ( σ 13 ) , gives the same information as previous column, but correlation was performed between sensors 1 and 3. r 14 ( σ 14 ) , In the same way as the previous two, brings the information regarding correlation between sensor 1 and 4. Finally, r 23 ( σ 23 ) shows behavior of the combination thumb-index (for exercise 3, it brings information regarding tweezers manipulation).
Exercise 3
Vol r 12 ( σ 12 ) r 13 ( σ 13 ) r 14 ( σ 14 ) r 23 ( σ 23 ) r 34 ( σ 34 ) r 24 ( σ 24 )
10.90 (0.15)0.94 (0.10)0.94 (0.10)0.89 (0.16)0.92 (0.17)0.90 (0.19)
20.88 (0.17)0.92 (0.12)0.91 (0.14)0.87 (0.18)0.94 (0.18)0.85 (0.20)
30.81 (0.10)0.97 (0.04)0.96 (0.05)0.80 (0.11)0.93 (0.13)0.83 (0.15)
40.87 (0.18)0.91 (0.16)0.92 (0.12)0.85 (0.19)0.90 (0.14)0.82 (0.19)
50.80 (0.19)0.90 (0.18)0.90 (0.17)0.78 (0.22)0.89 (0.15)0.80 (0.21)
60.80 (0.20)0.92 (0.19)0.91 (0.16)0.78 (0.23)0.92 (0.17)0.81 (0.21)
70.88 (0.16)0.91 (0.13)0.91 (0.16)0.86 (0.20)0.89 (0.16)0.84 (0.23)
80.83 (0.21)0.88 (0.15)0.86 (0.19)0.81 (0.26)0.91 (0.15)0.86 (0.22)
90.91 (0.11)0.97 (0.03)0.98 (0.05)0.89 (0.12)0.93 (0.09)0.91 (0.11)
100.80 (0.16)0.84 (0.27)0.89 (0.21)0.78 (0.30)0.88 (0.21)0.79 (0.24)
110.90 (0.18)0.91 (0.15)0.91 (0.15)0.88 (0.20)0.89 (0.17)0.90 (0.21)
120.89 (0.18)0.95 (0.04)0.95 (0.04)0.87 (0.18)0.91 (0.13)0.85 (0.19)
130.85 (0.10)0.96 (0.06)0.95 (0.12)0.85 (0.10)0.92 (0.11)0.87 (0.10)
140.84 (0.16)0.89 (0.14)0.87 (0.21)0.83 (0.20)0.90 (0.18)0.87 (0.21)
Table 8. r calculated between the sensor 1 and sensors 2, 3 and 4; and between sensors 2 and 3, during exercises 1 and 2. Columns in this table provide the same information shown in Table 7, but in this case for exercises 1 and 2.
Table 8. r calculated between the sensor 1 and sensors 2, 3 and 4; and between sensors 2 and 3, during exercises 1 and 2. Columns in this table provide the same information shown in Table 7, but in this case for exercises 1 and 2.
Exercise 1Exercise 2
Vol r 12 ( σ 12 ) r 13 ( σ 13 ) r 14 ( σ 14 ) r 23 ( σ 23 ) r 12 ( σ 12 ) r 13 ( σ 13 ) r 14 ( σ 14 ) r 23 ( σ 23 )
10.91(0.12)0.91(0.12)0.90(0.17)0.89(0.14)0.60(0.24)0.64(0.26)0.55(0.24)0.59(0.28)
20.92(0.10)0.93(0.09)0.91(0.12)0.90(0.11)0.75(0.13)0.75(0.12)0.76(0.13)0.74(0.15)
30.84(0.30)0.81(0.42)0.86(0.18)0.79(0.44)0.76(0.18)0.77(0.19)0.80(0.17)0.75(0.22)
40.80(0.13)0.82(0.13)0.81(0.14)0.79(0.14)0.67(0.18)0.62(0.20)0.69(0.18)0.61(0.25)
50.85(0.15)0.87(0.14)0.84(0.17)0.83(0.16)0.75(0.15)0.80(0.12)0.72(0.20)0.73(0.15)
60.90(0.11)0.91(0.11)0.90(0.11)0.89(0.12)0.67(0.26)0.65(0.26)0.67(0.24)0.65(0.26)
70.87(0.15)0.88(0.14)0.86(0.15)0.86(0.16)0.73(0.21)0.74(0.18)0.70(0.18)0.72(0.24)
80.71(0.19)0.69(0.22)0.69(0.18)0.67(0.25)0.78(0.13)0.86(0.18)0.86(0.09)0.78(0.23)
90.91(0.12)0.90(0.16)0.92(0.10)0.88(0.17)0.77(0.12)0.86(0.14)0.85(0.14)0.76(0.15)
100.85(0.21)0.86(0.21)0.87(0.20)0.83(0.26)0.67(0.21)0.69(0.18)0.49(0.30)0.66(0.26)
110.83(0.16)0.86(0.17)0.80(0.13)0.81(0.18)0.71(0.17)0.71(0.17)0.70(0.21)0.70(0.18)
120.92(0.08)0.94(0.06)0.91(0.08)0.89(0.12)0.67(0.18)0.75(0.12)0.68(0.22)0.66(0.20)
130.89(0.17)0.95(0.04)0.93(0.10)0.87(0.20)0.72(0.26)0.73(0.22)0.71(0.26)0.70(0.34)
140.85(0.14)0.86(0.15)0.89(0.11)0.85(0.17)0.65(0.24)0.66(0.22)0.64(0.21)0.63(0.26)
Table 9. Comparison of the results of groups Type 1 and Type 2. Each column represents the average and standard deviations for each variable on both groups. Type 1, has individuals with higher dexterity, as they used less time to complete the exercise, made fewer movements, and had smaller displacements. Type 2, include volunteers who have lower dexterity as the values of parameters were higher.
Table 9. Comparison of the results of groups Type 1 and Type 2. Each column represents the average and standard deviations for each variable on both groups. Type 1, has individuals with higher dexterity, as they used less time to complete the exercise, made fewer movements, and had smaller displacements. Type 2, include volunteers who have lower dexterity as the values of parameters were higher.
Group D u r a t i o n ¯   ( σ D u r a t i o n ) M o v e m e n t s ¯   ( σ M o v e m e n t s ) D i s p ¯   ( σ D i s p ) D i s p θ ¯   ( σ D i s p θ )
Type 145.87 (6.08)31.67 (3.88)2.24 (0.27)1099.17 (321.51)
Type 273.70 (15.05)53.38 (10.41)3.31 (0.95)1732.75 (440.78)
Table 10. Results from Kruskal-Wallis test for each exercise. In the case of exercises 1 and 2, results of the statistical test show that the IGlove responds in the same way in the case of exercises that are not suitable to assess dexterity. By contrast, the low ρ v a l u e s found for exercise 3 indicate a significant difference between groups, therefore, variables Duration, Movements, Disp and D i s p θ , acquired during FMS test can be used to assess dexterity using the IGlove.
Table 10. Results from Kruskal-Wallis test for each exercise. In the case of exercises 1 and 2, results of the statistical test show that the IGlove responds in the same way in the case of exercises that are not suitable to assess dexterity. By contrast, the low ρ v a l u e s found for exercise 3 indicate a significant difference between groups, therefore, variables Duration, Movements, Disp and D i s p θ , acquired during FMS test can be used to assess dexterity using the IGlove.
ρ V a l u e s
ExerciseDurationMovementsDisp D i s p θ
Exercise 10.60560.37050.89730.1967
Exercise 20.60560.74200.60560.6510
Exercise 30.00190.00290.00980.0098

Share and Cite

MDPI and ACS Style

Lemos, J.D.; Hernandez, A.M.; Soto-Romero, G. An Instrumented Glove to Assess Manual Dexterity in Simulation-Based Neurosurgical Education. Sensors 2017, 17, 988. https://doi.org/10.3390/s17050988

AMA Style

Lemos JD, Hernandez AM, Soto-Romero G. An Instrumented Glove to Assess Manual Dexterity in Simulation-Based Neurosurgical Education. Sensors. 2017; 17(5):988. https://doi.org/10.3390/s17050988

Chicago/Turabian Style

Lemos, Juan Diego, Alher Mauricio Hernandez, and Georges Soto-Romero. 2017. "An Instrumented Glove to Assess Manual Dexterity in Simulation-Based Neurosurgical Education" Sensors 17, no. 5: 988. https://doi.org/10.3390/s17050988

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop