Next Article in Journal
A Deep Learning-Based System for Monitoring the Number and Height Growth Rates of Moso Bamboo Shoots
Next Article in Special Issue
Specific Test Design for the In-Depth Technique Analysis of Elite Karate Competitors with the Application of Kinematic Sensors
Previous Article in Journal
Incorporating a Topic Model into a Hypergraph Neural Network for Searching-Scenario Oriented Recommendations
Previous Article in Special Issue
The Effect of 12 Weeks of Saddle Horse Conversion Training on Thoroughbred Horse Gait
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reducing the Number of Sensors in the Data Glove for Recognition of Static Hand Gestures

Institute of Electronics, Lodz University of Technology, 90-924 Lodz, Poland
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(15), 7388; https://doi.org/10.3390/app12157388
Submission received: 29 June 2022 / Revised: 17 July 2022 / Accepted: 18 July 2022 / Published: 22 July 2022
(This article belongs to the Special Issue Applied Biomechanics and Motion Analysis)

Abstract

:
Data glove devices, apart from being widely used in industry and entertainment, can also serve as a means for communication with the environment. This is possible thanks to the advancement in electronic technology and machine learning algorithms. In this paper, the results of the study using a designed data glove equipped with 10 piezoelectric sensors are reported, and the designed glove is validated on a recognition task of hand gestures based on 16 static signs of the Polish Sign Language (PSL) alphabet. The main result of the study is that recognition of 16 PSL static gestures is possible with a reduced number of piezoelectric sensors. This result has been achieved by applying the decision tree classifier that can rank the importance of the sensors for the recognition performance. Other machine learning algorithms were also tested, and it was showed that for the Support Vector Machines, k-NN and Bagged Trees classifiers, a recognition rate of the signs exceeding 90% can be achieved just for three preselected sensors. Such a result is important for a reduction in design complexity and costs of such a data glove with sustained reliability of the device.

1. Introduction

In the last decade, one can observe a significant development of virtual reality (VR) technology which has resulted in the introduction of systems using this technology to the consumer market. Such systems usually consist of a head-mounted display (HMD), commonly called VR goggles, controllers, and a pair of transceivers. The HMD is a screen, placed at such a distance from the eyes to create the impression of being in the displayed image. This impression is so strong that it has become known as the second reality—a virtual reality. However, a device that creates a near-reality phenomenon should also be expected to interact to further enhance the sense of real-world experience. Currently, the most used human interaction devices are joysticks called wands. They are a kind of combination of the human hand’s natural way of interacting with the world around us with existing input devices such as joysticks or computer game pads. For most people, especially youngsters, the control with such a “wand” is widely accepted and in a way natural, because they have been accustomed to using such devices since their childhood [1]. Additionally, the technologies for intuitive interaction with computers have been developed for years, e.g., through computer games. Computer games often require the player to perform several actions at the same time, e.g., moving simultaneously in the game space, controlling the camera, and performing a certain action. Additionally, online games require all these actions to be performed under stress due to action of the competing players and time pressure.
To identify the next stage in the development of human–computer interaction, one can follow the simplified path by which control of modern devices occurs. In order to perform the user’s original intention, whether it be a movement in a computer game or the desire to write on the computer keyboard, the user’s brain must first translate it into a series of hand movements and then enforce those movements. Thus, any interaction with a computer can be represented as a brain–hand–controller–computer chain [2].
There is already another solution to the problem of controlling a computer, which is the brain computer interface (BCI) [3]. However, it is currently an early stage technology, allowing for the simplest of actions, and is so expensive and non-ergonomic that it will not exist as a commercial solution for several years yet. It can therefore be concluded that the next significant step in the development of control technology will be to eliminate the need to use controller buttons. The greatest benefits of control with a controller that replicates the natural movements of our hand can be achieved when applied to objects with a range of motion similar to that of humans.
In the era of growing importance of telecontrol and breakthroughs in robotics, it also becomes necessary to develop a device that would allow an intuitive control of robotic manipulators while maintaining the required precision of their operation. However, for this to be possible, it is necessary to recreate the motor-nervous system of the hand so that the operator can freely perform operations and experience their consequences. This is possible with the use of devices that can track the user’s hand movements. A device that allows direct measurement of finger movements was developed and described by Sandin et al. [4] and was named data glove. The form of the glove makes the use of this device ergonomic and does not interfere with the tested movements. This allows for natural movement, which is very important when testing all kinds of gestures and sign language.
This paper describes a study using a designed data glove. The research focuses on two main objectives:
  • To test the effectiveness of the constructed glove in recognizing static hand gestures;
  • To investigate the possibility of reducing the number of sensors used in the recognition static hand gestures.
To the best of our knowledge, no research has yet been conducted on the feasibility of reducing the number of detectors in structures of similar purpose, and no research with publicly available results has been conducted on the Polish Sign Language alphabet using 10 flex sensors.
Reducing the number of sensors, and thus the size of the feature vector, will simplify not only the construction of the glove itself but also the classification process. A simpler design and a reduction in the necessary computational power could improve the usability of such a data glove. Furthermore, less complex data gloves that can be used as sign-language-to-speech translators may be less expensive and may be more widely accepted.

Sign Language Alphabet and Stance of Sign Language Speakers

It is important to draw a distinction between sign language as such and fingerspelling. Every sign language has its own grammar and vocabulary. However, the syntax of many of them is based on signs that are equivalents of whole words or phrases in a spoken language. The fingerspelling alphabet, on which the Polish version of the gestures presented in this article are based, is a supplement to the sign language. Fingerspelling is also used to make signs corresponding to counters, numbers, or punctuation. These sign language signs are made solely with the hands, two for British Sign Language and one for Polish Sign Language, among others, and do not require facial expressions or poses. Spelling in sign languages is most used when pronouncing proper names such as first names or street names or when learning other sign language signs [5].
The stance of the sign language community towards sign-language-to-speech glove-translators should also be made clear here. Many of them have a negative attitude towards this type of device. This is mainly due to bulkiness and low efficiency in sign language recognition [6]. While the first of these obstacles can be overcome by striving for miniaturization of such devices, for example, by reducing the number of sensors used (as described in this paper), the low effectiveness is due to the inability of the gloves to capture other important factors of sign language such as facial expression or pose [7]. However, although the data glove may not be the ultimate solution to the problem of translating sign language into speech, mainly due to the reasons stated above, it seems it is currently an excellent device for recognizing hand movements. The readings obtained with them and the databases created from them will allow for the development of more complete and accurate training sets for sign and sign word recognition algorithms, thus increasing their effectiveness in future implementations.
It should also be noted that the gestures referred in this work are not letters of the Polish Sign Language manual alphabet. It was assumed that the main difference between letters is their main, static form. However, such an assumption makes it necessary to classify the handshapes as gestures rather than letters of PSL due to lack of dynamic stages of presentation of individual characters.
The next section describes articles on applications of data gloves in gesture recognition and sign languages. Section 3 describes the design of the glove and the research methodology, and the results are presented in Section 4. The article concludes with a summary of the results in Section 5 and a brief discussion undertaken in Section 6.

2. Related Work

Analyzing the human hand as a mechanical system, there are a total of 27 degrees of freedom [8]. Considering individual human motoric coupling, the human hand is a very complex mechanical manipulator. In addition to the complex muscular and skeletal system, the hand (apart from the central nervous system), is the most innervated part of the human body with a significant cortical representation. The human brain devotes considerable resources to processing the signals of the sense of touch. This is a crucial and complex task, to the extent that it requires the synchronous use of both hemispheres of the brain. This is due to the need to provide feedback for all hand actions [9].
In [10], the authors reviewed numerous articles published in the past year about the data glove, focusing on their capabilities in recognizing sign language (SL) signs or gestures. It is important to point out the differences between SL signs, which can be divided in terms of the dynamics of a given gesture or sign. Static signs are presented in a sequence, and there is a natural suspension point in each sign. Dynamic gestures, on the other hand, are challenging due to the lack of such a point. In addition, with SL, quite like speech, the spacing between each sign is important so that it can be understood. Differences also occur at the level of semantics. In the case of SL, all its signs have a separate and independent meaning. Gestures can also be an independent form of communication, but there is a specific group of them whose role is only to reinforce another form of communication. Additionally, there is a group of gestures that have a purely mechanical function.
The documented designs of the data gloves can be divided into two main groups according to the principle of measurement (see Figure 1). The devices with indirect measurement method use a medium to read the position of the hand, and devices using the direct method are attached to it and thus track its movements.
Indirect measurement methods include electromagnetic waves and sound waves [8,9,11]. The most popular indirect method is the one that uses cameras to track a hand using machine vision techniques. A popular device of this type is Ultraleap’s Leap Motion, recently used to recognize 100 Arabic Sign Language gestures [11]. However, image-based hand tracking is subject to limitations due to the measurement method itself. The camera cannot capture movements that take place outside its field of view and those that are obscured by another object. Many everyday gestures or hand movements such as shaking hands may cause the skeleton tracking algorithm to be unable to map the position of the hand. This problem does not occur with hand tracking using the data glove [12,13].
The second group of devices are those that use sensors placed directly on the hand. These devices are called data gloves because of their glove-like shape. They can be classified according to the type of sensors used. One of the most popular types of sensors, probably due to its relatively low price and ease of application, are piezoresistive sensors [10].
Maitre et al. [14], using a self-built glove, investigated the behavior of the hand while holding seven everyday objects in the palm of the hand and made a successful attempt to classify these grasps based on the collected data.
Pezzouli et al. [15] conducted an experiment using a similar glove and tested the effectiveness of the device in recognizing 27 gestures and compared several methods for their segmentation and classification. Each glove has 10 deflection sensors and one Inertial Measurement Unit (IMU) for tracking hand orientation. The authors also conducted an experiment using a publicly available database containing readings obtained from a similar glove while performing sign language gestures. In both experiments, the team obtained high classification accuracy of 90% for most of the tested methods.
Mummadi et al. [16], on the other hand, presented a comparison of several methods for recognizing 22 static French sign language gestures. Data obtained from a built glove, employing five IMUs, were classified with the use of algorithms based on Random Forests, Naive Bayes Classifier and Multi-Layer Perceptron.
Zhang et al. [17] proposed a glove that is also based on piezoresistive sensors but in combination with an electromyograph and inertial sensors placed on the glove. The effectiveness of the solution was tested on 10 gestures, of which 5 were dynamic gestures, and achieved a recognition rate of 89.3% for static gestures and 76.6% for dynamic gestures.
Significant advances have also been made in the field of sensors themselves. Xie et al. [18] have developed a method for fabricating triboelectric nanogenerators for textronic applications. The nanogenerators are highly flexible and can be freely trimmed. Their electrical properties allow them to be used as active deflection sensors. Such a potential application was tested by measuring the voltage signals generated by the sensors when individual fingers were bent. A preliminary analysis of the obtained results shows the potential validity of this solution in data glove-type devices. A similar method and results were also reported in [19].
Huang et al. [20] presented sensors using graphene technology. A layer of reduced graphene oxide (RGO) was sputtered onto a piece of double-covered yarn (DCY) giving it piezoresistive properties. The performance of the sensors was tested on a set of 10 gestures representing digits from 0 to 9 and 9 dynamic Chinese Sign Language characters using artificial neural networks (ANN) [21] and dynamic time warping (DTW) method. The achieved recognition performance was 98.3%.
A different method of tracking finger movement was presented by Rinalduzzi et al. [22]. Coils were mounted on the fingertips of the glove, emitting an electromagnetic field at a different frequency for each of the seven transmitters. To function, such a solution needs receivers placed in three planes. The developers presented a wall made for this purpose, in whose space a gloved hand moves. This approach precludes the use of the device in open spaces which is one of the main drawbacks of using vision-based methods for gesture recognition.
In [23], a method of applying the sensors directly to the surface of the material is presented. The resistive fingertip deflection sensors on which the glove performance is based were fabricated using the direct indium–gallium eutectic writing method.
The software part of gesture recognition problem can be divided into two parts: detection and classification. While neural networks and machine learning algorithms can often achieve high classification performance, the need to capture the beginning and end moments of a given gesture poses a significant problem for devices serving as SL translators. Many systems developed to date have not been able to recognize dynamic SL characters, and none have operated in real time outside of laboratory environment. This is due to the difficulty of segmenting readings from such a glove. Separating sign language characters from each other is still a difficult challenge for the currently developed algorithms [10].
An original segmentation approach that is based on a scalar called Gesture Progression Scalar (GPS) was proposed by Lee and Bae [23]. The GPS is determined by the dynamics of resistance changes in 10 sensors. Based on its values, it is possible to determine the moment at which a given gesture starts and ends. It is also worth mentioning that both presented algorithms were tested on a dedicated glove without any inertial sensor, which is a typical approach for dynamic gesture recognition. The ANN-based recognition algorithm was tested on 10 American SL characters.
The final step in the gesture recognition pipeline is gesture classification. Machine learning algorithms are often used for gesture classification. Their use allows researchers to obtain fast classification at the cost of longer learning times [10]. ANN was used by Lee and Bae [23] and Huang et al. [20]; machine learning algorithms such as Support Vector Machine [10,18,19,20,21,22,23,24] or k-Nearest Neighbors [21,22,23,24,25,26] and Hidden Markov Models [27,28,29,30] are equally often used and juxtaposed.

3. Materials and Methods

The glove was made of materials and components generally available on the consumer market. The main objectives of the design were to allow free movement of the sensors while maintaining their position in relation to the finger and the possibility of adjusting the glove to the hand of the subjects. Due to the phenomenon of cracking of the connectors of the sensors used, they were replaced with crimped Amphenol FCI connectors, which additionally enabled quick replacement of most sensors in case of damage or wear.

3.1. Hardware

The glove used in the experiment was made of 10 piezoresistive flex sensors: 4 with a length of 112 mm and 6 shorter ones with a length of 73 mm. They were placed on the glove in two rows on each of the five fingers: the first at the level of metacarpophalangeal joints and the second covering interphalangeal joints. The choice of sensor placement assumed that independent flexion of the distal interphalangeal (DIP) joints does not occur in sign languages and that it does not belong to any group of gestures used in everyday life. Additionally, only a small number of people can perform such a flexion naturally.
The sensors were attached to the textile surface of the glove using Velcro. The motion path of the sensors was stabilized by 3D-printed guides which are also attached to the glove using Velcro. Movable sensor mounts allowed us to adjust their placement and motion path to the user’s finger spacing. In this way, the axis of the sensor always coincided with the finger axis during measurements, making them as accurate as possible. A photograph of the glove is shown in Figure 2. The sensors were connected to the measurement unit using conductive threads embedded in the glove. The ends of the thread were attached to the sensors with detachable connectors so that the sensors could be easily replaced. The other end of the thread was soldered to the PCB of the measurement system. On the PCB, there are also 15 kΩ resistors to which piezoresistive sensors are soldered so creating a voltage divider.
Linearity of the resistance characteristics of the sensor with respect to the bending angle was confirmed with a simple experiment measuring the resistance of the sensor in the bending range from 0° to 90° in 5° steps; it was carried out using Axiomet AX-8455 multimeter manufactured by Shanghai Yi Hua V&A Instrument Co. in Shanghai, China.
The measurements of glove sensors are controlled by a microcomputer Raspberry Pi Zero W, which is a single-board computer running on a Broadcom BCM2835 1 GHz processor with 512 MB of built-in RAM. It consists of 4 ADS1115 16-bit analog-to-digital converters working at 860 Hz sampling rate (see Figure 3).
The operation of the glove was controlled by the Linux operating system in a distribution designed for embedded systems with the program controlling the operation of the acquisition system written in Python.

3.2. Database Generation

A special measurement procedure was applied to collect the measurement data from the data glove. The script implementing this procedure controls the readings of 10 sensors in a continuous mode, enabling quick diagnostics of the sensors’ operation. After a subject assumed the final, static position of a gesture shown in Figure 4, the script saves 10 consecutive readings of sensor values to a csv file that was automatically created at the start of the measurement.
There were 15 subjects aged from 23 to 68 who volunteered in the experiments. Five women and ten men took part in the trials. Three of the subjects were left-handed. All subjects had an unrestricted range of hand movement necessary for the study. They had similar hand size ranging from 18 to 21 cm measured from the wrist to the tip of the middle finger, with no known musculoskeletal disorders. None of the subjects had also been exposed to sign language prior to the study.
The tests of the data glove were approved by the Institutional Ethics Committee of Technical University of Lodz (No. 1/2021, date of approval: 28 October 2021). Before starting the measurement, subjects were instructed about the course of the measurement (shown in Figure 5) and were familiarized with 16 gestures based on letters of the Polish Sign Language fingerspelling alphabet, which they were about to present. Each participant performed each of 16 static gestures 10 times. After each repetition, the hand was relaxed for a few seconds, which resulted in slight differences between repetitions.
The measurements of each subject were taken in a natural sitting position, with the elbow resting on a surface and the forearm gently deviating from the vertical. Finger placement in the natural relaxed position was also measured in this position. In total, the database consisted of 25,550 readings of the sensors.
The nominal resistances of the sensors differed even relative to sensors from the same production batch. The reading distribution chart shown in Figure 6 presents the median readings and outliers. The formation of outliers was caused by an imperfection in the way the sensors were connected to the conductive threads unnoticeable at the time of the study. This minor flaw was corrected, while the outliers remained in the dataset on purpose, reflecting an imperfection in the actual functioning of the device. In all the presented results, we used the data with outliers for training the classifiers.

4. Results

Of the many existing sign languages, it was decided to conduct the research based on the Polish Sign Language. In the case of static characters, it is not important which hand they are made with. Therefore, for practical reasons, the left hand was chosen.
All classification processes were performed using a five-fold cross-validation technique.
For the standard k-NN algorithm methods, the k parameter was set to 1. For the k-NN in Random Spaces (k-NNiRS) method, the number of subspaces was chosen automatically by the software and was equal to about half of the total number of data dimensions.
The goal of the experiment was to investigate the possibility of reducing the number of sensors while losing as little gesture recognition performance as possible.
For this purpose, the Maximum Relevance-Minimum Redundancy (MRMR) method was used. It is an iterative method of determining the most relevant features of the dataset under study. The score for each variable is determined as the quotient of significance and redundancy against previously tested variables. Relevance is calculated using F-statistic between the feature and the expected class, while redundancy is calculated as the average Pearson correlation between the variable under study and the variables checked in previous iterations [31].
Once the sensor hierarchy presented in Figure 7 was determined, data classification was performed using six classifiers. Almost all relevance metrics identified sensors P3_2, P1_2 and P2_2 as the three most relevant. The least significant sensor according to all selection methods was P4_2.
The effectiveness of each algorithm was calculated using Equation (1)
Performance = TPR TPR + FPR × 100 %
where TPR means True Positive Rate and FPR stands for False Positive Rate.
The effectiveness of each algorithm for a limited number of sensors is shown in Figure 8.
The average gesture recognition performance of each tested algorithm from all conducted experiments is shown in Table 1.
The highest level of correct classifications was achieved by the k-NNiRS classifier when working on all 10 sensors. A slightly lower recognition rate was obtained for the Bagged Trees method. The BT method, in contrast to k-NNiRS, maintains a performance above 94% regardless of the number of sensors and their placement (Figure 8 and Figure 9). To verify how reducing the number of dimensions in the feature vector affects the efficiency of distinguishing individual characters, the confusion matrix was built for the algorithm with the highest mean efficiency.
Additionally, we re-run the classification process ignoring data from the first interphalangeal (IP) and then the second metacarpophalangeal (MCP) row of sensors to check how placement over specific joints impacts the recognition rates.
The obtained results allow us to conclude that sensor placement over specific joints does not affect the performance of most classifiers. For the Classification Tree and Naive Bayes Classifier, the placement on the metacarpophalangeal joints achieving a 20% improvement in recognition rate.
For the three data classifiers with the highest average performance, the dependence of classification performance on the dimensionality of the feature vector was also tested. In this case, components of the feature vector were sequentially included into the classification process starting with those with the highest importance index as determined by the MRMR method.
In terms of the k-NNiRS method, for the full ten-component feature vector, a gesture recognition rate of 98.9% was achieved, while for the three-component feature vector, the recognition rate dropped to −89.1%. As the dimensionality of the dataset decreased, there was a decrease in recognition performance for all characters. However, a several-fold increase in False Negative Rates for a few characters, i.e., “W”, “O”, “U”, “R” and the resting state was observed. The largest drop in recognition performance for the W”, “O” and “U” gestures is most likely due to the exclusion of the sensor at the thumb’s interphalangeal joint from the classification process. Without this information, the algorithm failed to distinguish whether the thumb was fully flexed or whether the only flexion occurred at the interphalangeal joint and the thumb remains straight. The decrease in performance in recognizing a gesture based on the letter “R” is due to the disabling of the P2_2 sensor at the interphalangeal joint of the index finger. Even though the range of motion necessary to record movement of the fingers at the metacarpophalangeal joint bringing the fingers closer and further apart is theoretically unattainable for the sensors used placed in this plane, the P2_2 sensor nevertheless recorded the overlap of the index and middle fingers that make up the “R” sign. The most poorly recognized sign using a three-component feature vector is the sign based on the letter “W.” The cause of this is the similarity of this gesture to many others in the form of straightened fingers 2–4 and a bent thumb and little finger. In addition, for many people, this positioning of the fingers is extremely unnatural, which not always resulted in the correct performance of this gesture. Table 2 shows the confusion matrix obtained for character classification using the most effective k-NNiRS algorithm, with a reduced number of sensors used.
To test the recognition performance of characters not relying on the P3_2, P1_2 and P2_2 sensors, the average readings for two other very similar characters were examined. The characters “O” and “S” differ in the flexion of only one joint. As can be seen in Figure 10, the average readings for the two characters clearly point to the P1_1 sensor taking different readings.

5. Discussion

An interesting feature of the PSL is that it has diacritics that are dynamic versions of corresponding letters. This feature is important for further research involving recognition of dynamic signs.
The initial phase of the study focused on testing the overall performance of the data glove. A significant challenge was the recognition of gestures with similar finger arrangements. As shown in Figure 10, the differences in two similar gestures captured by the presented glove are visible to the naked eye. The finger placements for both gestures are very similar to each other and differ only in the flexion of one joint.
Importantly, reducing the dimensionality of the feature vector according to the designated sensor importance did not result in a significant decrease in the classification performance of the recognized characters (Figure 11).
Reducing the dimensionality to five features resulted in a 97.5% effectiveness for k-NNiRS and BT algorithms.
Although the efficiency of the k-NNiRS algorithm turned out to be the highest of all classifier for 10 sensors, it decreased most rapidly as the dimensionality of the feature vector was reduced. This is due to the main assumption of the algorithm, which includes self-reduction method of the feature vector to a randomly selected few of its components. This explains the apparent decrease in classification efficiency of the k-NNiRS algorithm, as seen in Figure 11, while the BT and k-NN algorithms achieve efficiency of 85%. The k-NNiRS algorithm in this situation uses only a one-dimensional feature vector.
Table 3 compares the work presented in Related Work with the results presented in this paper. It should be noted that the solutions for recognizing dynamic gestures and static gestures cannot be directly compared due to the different methods used. The main contribution of study is the method of reducing the number of sensors needed to recognize static gestures based on Polish Sign Language to 3 out of 10 sensors. As can be seen in Table 3, no device presented in the previous works used a device with such a small number of sensors.
An efficiency of 95.1% was achieved by BT and k-NN classifiers for the three-component feature vector (Figure 8). A more than ten-fold increase in FPR was observed for several hand positions including those corresponding to letters “W”, “O”, “U” and “R”. The gestures corresponding to letters “W” and “U” were most often confused with the gesture corresponding to letter “R”. This is most likely because all these gestures require the index and middle finger to be fully straightened, while overlapping these fingers while performing the gesture corresponding to letter “R” requires detection of a movement that is not performed in the plane of movement of the sensor. The decrease in TPR for the letter “O” can be explained in a similar way. It relies mainly on the index finger and thumb, with middle finger straight.
The exclusion from the classification process of the sensor located above the interphalangeal joints of the index finger, which flexes the most during this gesture, may result in a false prediction of the gesture. Nevertheless, the performance obtained can be considered satisfactory and promising, as it remains above or close to 90% for most gestures. This is a result worth noting because k-NN and BT are classifiers with short training time and low computational complexity.
The proposed reduction in the number of sensors results in a decrease in the recognition efficiency of the tested gestures to 93–95%. Table 2 shows that the largest contribution to these errors comes from misrecognized gestures corresponding to letters “W” and “U”.
In its current form, the system can be used for simple fingerspelling interactions. It would only allow the user to provide simple information or to translate proper names to people without knowledge of sign language or write on a computer using the manual alphabet. A version of the system having a set of 10 flex sensors and inertial sensors will also be an excellent tool for creating freely available databases. The aim of a future study is to develop a system consisting of two hand overlays based on as few sensors as possible. A solution could be a system that combines hand motion recognition using a glove and a smartphone camera to track mouth movements, pose, and facial expressions of the speaker.

6. Conclusions

For the recognition of hand gestures drawn from PSL static letters, it is not necessary to use 10 sensors monitoring the flexion of all members of each of the 5 fingers to achieve classification performance exceeding 97%.
For the problem of recognizing the tested static hand gestures, three sensors at the level of the metacarpal joints are most relevant: on the middle finger, on the thumb, and on the index finger.
The classification process based only on these three values achieves a performance of 95% for the k-NN method for k = 1. It is also the most resilient to reductions in the dimensionality of the input data and to interference from hardware defects.
Despite the exclusion of key features for some characters, the recognition rate for most characters is sustained at a level close to 90%.
The above conclusions apply only to the studied case of the static gestures. The next step in the research will be to test the validity of reducing the number of sensors in the dynamic hand gestures of an actual PSL manual alphabet recognition task, also performed in sequence. Further studies will solicit the participation of PSL users when the feasibility of using fewer sensors in recognizing actual PSL letters is confirmed by preliminary results. With their participation, the plan is to continue building a database of glove signals when using PSL. If the above proposal is positively received by the sign language community, the possibility of reducing the number of sensors in data glove-type devices used as sign language translators can bring significant benefits to both designers and potential users. We believe that using fewer sensors would not only simplify the design and thus reduce the number of unreliable components and the price of the device but would also simplify and speed up the sign recognition process.

Author Contributions

J.P.: conceptualization, software, validation, formal analysis, investigation, resources, data curation, writing—original draft preparation, visualization. P.S.: methodology, supervision, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Ethics Committee of Technical University of Lodz (1/2021, date of approval: 28 October 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this article are publicly available at: https://figshare.com/search?q=10.6084%2Fm9.figshare.20304132 (accessed on 13 July 2022).

Acknowledgments

Jakub Piskozub would like to thank Lodz University of Technology for granting a scholarship.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vogels, E.A. Millennials Stand Out for Their Technology Use|Pew Research Center, 9 August 2019. Available online: https://www.pewresearch.org/fact-tank/2019/09/09/us-generations-technology-use/ (accessed on 13 February 2021).
  2. Somatosensory Pathways (Section 2, Chapter 4) Neuroscience Online. An Electronic Textbook for the Neurosciences; Department of Neurobiology and Anatomy—The University of Texas Medical School: Houston, TX, USA, 2020; Available online: https://nba.uth.tmc.edu/neuroscience/m/s2/chapter04.html (accessed on 13 February 2021).
  3. Ortiz-Echeverri, C.J.; Salazar-Colores, S.; Rodríguez-Reséndiz, J.; Gómez-Loenzo, R.A. A new approach for motor imagery classification based on sorted blind source separation, continuous wavelet transform, and convolutional neural network. Sensors 2019, 19, 4541. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Sandin, D.J.; DeFanti, T.A. US NEA R60-34-163 Final Project Report; Sayre Glove: Online, 1977; pp. 1–30. Available online: https://www.evl.uic.edu/documents/us_nea_r60-34-163_sayre_glove_finalreport_november1977.pdf (accessed on 7 February 2021).
  5. Fingerspelling in American Sign Language|Start ASL. Available online: https://www.startasl.com/fingerspelling/ (accessed on 2 January 2022).
  6. Grieve-Smith, A. Ten Reasons Why Sign-To-Speech Is not Going to Be Practical Any Time Soon—Technology and Language. Available online: http://grieve-smith.com/blog/2016/04/ten-reasons-why-sign-to-speech-is-not-going-to-be-practical-any-time-soon/ (accessed on 14 December 2021).
  7. Why the Signing Gloves Hype Needs to Stop—Audio Accessibility. Available online: https://audio-accessibility.com/news/2016/05/signing-gloves-hype-needs-stop/ (accessed on 14 December 2021).
  8. ElKoura, G.; Singh, K. Handrix: Animating the Human Hand. In Proceedings of the 2003 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, San Diego, CA, USA, 26–27 July 2003. [Google Scholar]
  9. Gandhoke, G.S.; Belykh, E.; Zhao, X.; Leblanc, R.; Preul, M.C. Edwin Boldrey and Wilder Penfield’s Homunculus: A Life Given by Mrs. Cantlie (In and Out of Realism). World Neurosurg. 2019, 132, 377–388. [Google Scholar] [CrossRef] [PubMed]
  10. Ahmed, M.A.; Zaidan, B.B.; Zaidan, A.A.; Salih, M.M.; Bin Lakulu, M.M. A Review on Systems-Based Sensory Gloves for Sign Language Recognition State of the Art between 2007 and 2017. Sensors 2018, 18, 2208. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Deriche, M.; Aliyu, S.O.; Mohandes, M. An Intelligent Arabic Sign Language Recognition System Using a Pair of LMCs With GMM Based Classification. IEEE Sens. J. 2019, 19, 8067–8078. [Google Scholar] [CrossRef]
  12. Sensoryx. Available online: https://vimeo.com/user66039489 (accessed on 11 February 2021).
  13. Do, N.-T.; Kim, S.-H.; Yang, H.-J.; Lee, G.-S. Robust Hand Shape Features for Dynamic Hand Gesture Recognition Using Multi-Level Feature LSTM. Appl. Sci. 2020, 10, 6293. [Google Scholar] [CrossRef]
  14. Maitre, J.; Rendu, C.; Bouchard, K.; Bouchard, B.; Gaboury, S. Basic Daily Activity Recognition with a Data Glove. Procedia Comput. Sci. 2019, 151, 108–115. [Google Scholar] [CrossRef]
  15. Pezzuoli, F.; Corona, D.; Corradini, M.L. Recognition and Classification of Dynamic Hand Gestures by a Wearable Data-Glove. SN Comput. Sci. 2020, 2, 5. [Google Scholar] [CrossRef]
  16. Mummadi, C.K.; Leo, F.P.P.; Verma, K.D.; Kasireddy, S.; Scholl, P.M.; Kempfle, J.; Van Laerhoven, K. Real-Time and Embedded Detection of Hand Gestures with an IMU-Based Glove. Informatics 2018, 5, 28. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, X.; Yang, Z.; Chen, T.; Chen, D.; Huang, M.-C. Cooperative Sensing and Wearable Computing for Sequential Hand Gesture Recognition. IEEE Sens. J. 2019, 19, 5775–5783. [Google Scholar] [CrossRef]
  18. Xie, L.; Chen, X.; Wen, Z.; Yang, Y.; Shi, J.; Chen, C.; Peng, M.; Liu, Y.; Sun, X. Spiral Steel Wire Based Fiber-Shaped Stretchable and Tailorable Triboelectric Nanogenerator for Wearable Power Source and Active Gesture Sensor. Nano-Micro Lett. 2019, 11, 39. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Chiu, C.-M.; Chen, S.-W.; Pao, Y.-P.; Huang, M.-Z.; Chan, S.-W.; Lin, Z.-H. A smart glove with integrated triboelectric nanogenerator for self-powered gesture recognition and language expression. Sci. Technol. Adv. Mater. 2019, 20, 964–971. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Huang, X.; Wang, Q.; Zang, S.; Wan, J.; Yang, G.; Huang, Y.; Ren, X. Tracing the Motion of Finger Joints for Gesture Recognition via Sewing RGO-Coated Fibers Onto a Textile Glove. IEEE Sens. J. 2019, 19, 9504–9511. [Google Scholar] [CrossRef]
  21. Haykin, S. Neural Networks: A Comprehensive Foundation, 1st ed.; Prentice Hall PTR: Hoboken, NJ, USA, 1994. [Google Scholar]
  22. Rinalduzzi, M.; De Angelis, A.; Santoni, F.; Buchicchio, E.; Moschitta, A.; Carbone, P.; Bellitti, P.; Serpelloni, M. Gesture Recognition of Sign Language Alphabet Using a Magnetic Positioning System. Appl. Sci. 2021, 11, 5594. [Google Scholar] [CrossRef]
  23. Lee, M.; Bae, J. Deep Learning Based Real-Time Recognition of Dynamic Finger Gestures Using a Data Glove. IEEE Access 2020, 8, 219923–219933. [Google Scholar] [CrossRef]
  24. Shwetha, V.; Vijayalaxmi; Asarpota, D.A.; Verma, H. User trainable sign language to speech glove using KNN classifier. Compusoft 2019, 8, 3053–3058. [Google Scholar]
  25. Ha, N.; Withanachchi, G.P.; Yihun, Y. Force Myography Signal-Based Hand Gesture Classification for the Implementation of Real-Time Control System to a Prosthetic Hand. In Proceedings of the 2018 Design of Medical Devices Conference, Minneapolis, MN, USA, 9–12 April 2018; pp. 12–15. [Google Scholar] [CrossRef] [Green Version]
  26. Anh, L.H.; Quang, N.M.; Long, N.D.; Khoa, N.D.T.A.A.D.D. Implementation of data glove using 9-axis IMUS for gestures recognition and biomimetic robot hand control. In Proceedings of the 1st National Scientific Conference on Dynamics and Control, Da Nang, Vietnam, 19–20 July 2019; pp. 152–159. [Google Scholar] [CrossRef]
  27. Saggio, G.; Cavallo, P.; Ricci, M.; Errico, V.; Zea, J.; Benalcázar, M. Sign Language Recognition Using Wearable Electronics: Implementing k-Nearest Neighbors with Dynamic Time Warping and Convolutional Neural Network Algorithms. Sensors 2020, 20, 3879. [Google Scholar] [CrossRef] [PubMed]
  28. Pan, J.; Luo, Y.; Li, Y.; Tham, C.-K.; Heng, C.-H.; Thean, A.V.-Y. A Wireless Multi-Channel Capacitive Sensor System for Efficient Glove-Based Gesture Recognition With AI at the Edge. IEEE Trans. Circuits Syst. II Express Briefs 2020, 67, 1624–1628. [Google Scholar] [CrossRef]
  29. Pławiak, P.; Sośnicki, T.; Niedźwiecki, M.; Tabor, Z.; Rzecki, K. Hand Body Language Gesture Recognition Based on Signals From Specialized Glove and Machine Learning Algorithms. IEEE Trans. Ind. Inform. 2016, 12, 1104–1113. [Google Scholar] [CrossRef]
  30. Yuan, G.; Liu, X.; Yan, Q.; Qiao, S.; Wang, Z.; Yuan, L. Hand Gesture Recognition using Deep Feature Fusion Network based on Wearable Sensors. IEEE Sens. J. 2021, 21, 539–547. [Google Scholar] [CrossRef]
  31. Hassan, M.; Assaleh, K.; Shanableh, T. Multiple Proposals for Continuous Arabic Sign Language Recognition. Sens. Imaging 2019, 20, 4. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Classification of hand movement measuring devices according to their principle of operation.
Figure 1. Classification of hand movement measuring devices according to their principle of operation.
Applsci 12 07388 g001
Figure 2. The glove used in the study.
Figure 2. The glove used in the study.
Applsci 12 07388 g002
Figure 3. Block diagram of the glove.
Figure 3. Block diagram of the glove.
Applsci 12 07388 g003
Figure 4. Gestures based on static letters of Polish Sign Language alphabet.
Figure 4. Gestures based on static letters of Polish Sign Language alphabet.
Applsci 12 07388 g004
Figure 5. The methodology flowchart.
Figure 5. The methodology flowchart.
Applsci 12 07388 g005
Figure 6. Distribution of measurements data.
Figure 6. Distribution of measurements data.
Applsci 12 07388 g006
Figure 7. Importance of specific sensors.
Figure 7. Importance of specific sensors.
Applsci 12 07388 g007
Figure 8. Classification efficiency as a function of the number of sensors used.
Figure 8. Classification efficiency as a function of the number of sensors used.
Applsci 12 07388 g008
Figure 9. Classification performance as a function of the row of sensors used.
Figure 9. Classification performance as a function of the row of sensors used.
Applsci 12 07388 g009
Figure 10. The average signal values of each sensor for the gestures corresponding to letters “O” and “S”.
Figure 10. The average signal values of each sensor for the gestures corresponding to letters “O” and “S”.
Applsci 12 07388 g010
Figure 11. Graphs of gesture classification performance by algorithms with the highest average performance against the number of sensors.
Figure 11. Graphs of gesture classification performance by algorithms with the highest average performance against the number of sensors.
Applsci 12 07388 g011
Table 1. The average effectiveness of each of the 6 classifiers tested.
Table 1. The average effectiveness of each of the 6 classifiers tested.
AlgorithmAverage Recognition Rate [%]
Classification Tree76.7
Kernel Naive Bayes Classifier74.6
Cubic SVM91.5
k-NN93.7
k-NN in RS91.5
Bagged Trees93.4
Table 2. Confusion matrix for k-NNiRS method, while k = 1, obtained with feature vector limited to three sensors (expressed as a percentage).
Table 2. Confusion matrix for k-NNiRS method, while k = 1, obtained with feature vector limited to three sensors (expressed as a percentage).
TRUE CLASSa95.20.61.00.01.00.10.00.10.10.30.10.10.10.30.50.50.195.24.8
b0.490.80.90.60.10.50.30.10.60.51.10.60.81.40.90.00.490.89.2
c0.40.688.11.00.51.20.40.40.81.80.81.00.60.60.80.50.488.111.9
e0.10.80.592.50.10.60.61.20.20.40.21.40.60.50.20.20.192.57.5
i0.50.20.50.095.90.50.00.10.50.30.00.00.30.30.30.30.495.94.1
l0.40.41.11.00.692.20.10.60.11.30.10.10.20.30.11.00.592.27.8
m0.00.10.11.90.00.392.91.40.30.10.10.60.80.30.40.10.692.97.1
n0.00.40.41.60.00.21.987.81.70.60.52.20.90.80.50.10.587.812.2
o0.40.41.60.51.00.10.61.484.31.11.41.50.61.81.61.00.684.315.7
p0.50.41.40.80.11.80.30.40.789.60.80.70.60.40.70.30.589.610.4
r0.11.10.30.30.10.50.30.61.60.386.20.50.63.32.81.00.586.213.8
s0.40.60.51.40.10.30.51.42.10.60.289.61.40.50.30.10.189.610.4
t0.01.00.50.40.10.31.00.60.70.91.51.088.21.90.80.30.888.211.8
u0.11.40.80.30.50.10.51.01.40.12.80.71.483.82.42.10.683.816.2
w0.31.30.40.30.60.10.40.82.30.83.20.40.72.782.22.60.982.217.8
y0.90.10.60.01.11.20.10.30.70.80.70.10.31.71.689.30.589.310.7
rest0.31.00.80.20.20.30.70.91.01.21.00.31.60.92.01.785.685.614.4
abceilmnoprstuwyrestTPRFPR
PREDICTED CLASS
Table 3. Comparison of the results presented in this paper to articles in Related Work.
Table 3. Comparison of the results presented in this paper to articles in Related Work.
AuthorRef. Num.Kind of GesturesHighest PerformanceNumber of SensorsMethod# Gestures# Users
Maitre et al.[14]static, based on daily activities9915 piezoresistivek-NN/RF89
Pezzouli et al.[15]dynamic, based on Australian Sign Language99.710 piezoresistive + 1 IMURF275
Mummadi et al.[16]static, based on French Sign Language915 IMUsRF/ANN2217
Zhang et al.[17]dynamic, based on daily activities89.345 piezoresistive + Myo armbandLSTM1010
Xie et al.[18]static, based on daily activitiesn/a5 custom made spiral steel wire basedn/a81
Huang et al.[20]dynamic, based on Chinese Sign Language98.310 RGO piezoresistiveANN92
Lee and Bae[23]dynamic, based on American Sign Language10010 GaIn piezoresistiveLSTM111
This workn/astatic, based on Polish Sign Language98.9/94.510/3 piezoresistivek-NNiRS/
k-NN
1715
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Piskozub, J.; Strumillo, P. Reducing the Number of Sensors in the Data Glove for Recognition of Static Hand Gestures. Appl. Sci. 2022, 12, 7388. https://doi.org/10.3390/app12157388

AMA Style

Piskozub J, Strumillo P. Reducing the Number of Sensors in the Data Glove for Recognition of Static Hand Gestures. Applied Sciences. 2022; 12(15):7388. https://doi.org/10.3390/app12157388

Chicago/Turabian Style

Piskozub, Jakub, and Pawel Strumillo. 2022. "Reducing the Number of Sensors in the Data Glove for Recognition of Static Hand Gestures" Applied Sciences 12, no. 15: 7388. https://doi.org/10.3390/app12157388

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop