Next Article in Journal
Signal Processing for the Condition-Based Maintenance of Rotating Machines via Vibration Analysis: A Tutorial
Previous Article in Journal
Triboelectric Contact Localization Electronics: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

American Sign Language Recognition and Translation Using Perception Neuron Wearable Inertial Motion Capture System

1
Faculty of Informatics, Gunma University, Kiryu 3768515, Japan
2
Graduate School of Engineering, Hokkaido University, Sapporo 0608628, Japan
3
Faculty of Engineering, Hokkaido University, Sapporo 0608628, Japan
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(2), 453; https://doi.org/10.3390/s24020453
Submission received: 9 November 2023 / Revised: 17 December 2023 / Accepted: 10 January 2024 / Published: 11 January 2024
(This article belongs to the Section Wearables)

Abstract

:
Sign language is designed as a natural communication method to convey messages among the deaf community. In the study of sign language recognition through wearable sensors, the data sources are limited, and the data acquisition process is complex. This research aims to collect an American sign language dataset with a wearable inertial motion capture system and realize the recognition and end-to-end translation of sign language sentences with deep learning models. In this work, a dataset consisting of 300 commonly used sentences is gathered from 3 volunteers. In the design of the recognition network, the model mainly consists of three layers: convolutional neural network, bi-directional long short-term memory, and connectionist temporal classification. The model achieves accuracy rates of 99.07% in word-level evaluation and 97.34% in sentence-level evaluation. In the design of the translation network, the encoder-decoder structured model is mainly based on long short-term memory with global attention. The word error rate of end-to-end translation is 16.63%. The proposed method has the potential to recognize more sign language sentences with reliable inertial data from the device.

1. Introduction

Sign language is the primary communication method among deaf and hard-of-hearing people. According to the World Health Organization, nearly 2.5 billion people are projected to have some degree of hearing loss, and at least 700 million will require hearing rehabilitation by 2050 [1]. With the development of artificial intelligence technology, more and more deep learning models appear in natural language processing (NLP), bringing about significant changes in human society. Therefore, it is critical to apply deep learning models to promote the research on sign language translation.
Prior work of sign language recognition can be divided into the following two main mechanisms: vision-based and wearable sensors-based recognition. The vision-based approach utilizes RGB or RGB-D cameras to catch the dynamic movements of the hands [2,3]. Some of the commonly used corpora containing a large amount of data are RWTH-PHOENIX-Weather [4] and Chinese Sign Language (CSL) [5]. The RWTH-PHOENIX-Weather selects the sign language interpretation of daily news and weather forecasts over three years. The CSL dataset collected by Kinect contains 100 continuous Chinese sign language sentences.
Compared with vision-based methods, the data amount of wearable sensors-based corpora is generally small due to limited data sources. The MyoSign [6] covering 70 commonly used American sign language (ASL) words and 100 ASL sentences incorporated multimodal Convolutional Neural Network (CNN), bidirectional Long Short Term Memory (LSTM), and Connectionist Temporal Classification (CTC) to achieve continuous sign language recognition. The sentences were stitched together by the gestures of 70 words, but not coherent actions of sign language performance. The SignSpeaker [7] deployed on a smartwatch and a smartphone to realize real-time, robust, and user-friendly sign language recognition. One hundred and three common-used words covering the diversity of the ASL actions were selected to generate 73 sentences, following the grammar of ASL. WearSign [8] leveraged a smartwatch and an armband to capture sophisticated sign gestures. In addition, a multi-task encoder-decoder framework was introduced to realize the end-to-end translation of 250 ASL sentences.
Wearable devices commonly used for sign language recognition include independent Electromyography (EMG) & inertial measurement unit (IMU) sensors, armbands, and data gloves. The Myo armband is a lightweight off-the-shelf wearable device providing nine inertial and eight-channel EMG signals. Many studies on sign language word and gesture recognition have used this device considering the portability of the armband [9,10,11,12,13]. Savur & Sahin [14] classified the gestures of 26 English alphabet letters with Myo armband. Jane & Sasidhar [15] did the classification of 48 sign language words. Paudyal et al. [16] achieved an accuracy of 97.72% for 30 ASL gestures. In addition to the armband, independent EMG sensors have been used in various studies [17,18]. Yu et al. [19] applied four EMG sensors and one inertial sensor spreading around the forearm to recognize 150 CSL subwords by the deep belief net. With a fusion strategy for combining multi-sensor and multi-channel information, Li et al. [20] developed a 121 CSL subwords recognition system with two 3-dimensional accelerometers (ACC) and eight EMG sensors worn on the forearm. It is also a novel idea to customize data gloves [21,22]. Feng et al. [23] utilized novel triboelectric textile sensors attached to the glove to measure hand gestures. Then, the glove signals were used to recognized 50 words and 20 sentences with the deep learning model. Korzeniewska et al. [24] presented a data glove using textronic elements produced by a physical vacuum deposition process. The ASL alphabet recognition accuracy of this device was 86.5%. Other wearables which are also suitable for hand gesture recognition include smartwatches, smart rings, and acoustic devices. Liu et al. [25] combined a smartwatch worn on the wrist and a smart ring worn on the index finger to recognize the 100 most frequently used ASL finger gestures. The accuracy was 94.2%. In addition, SonicASL [26] leveraged dual speakers and microphones to capture the sonic feedback from hand gestures. Given 42 frequently used ASL words and 30 sentences, the system could achieve an accuracy of 93.8% at the word level, and 90.6% at the sentence level. The summary of previous studies is listed in Table 1.
In the above-mentioned studies on sign language recognition using wearable sensors, more studies focused on single gestures. Few datasets for sentence recognition were confined to a limited number of sentences and did not consider grammatical inconsistencies between sign language and spoken language. Datasets containing more sentences for the end-to-end translation had a limited number of words in the vocabulary. The selected words restricted more natural expressions. Based on the experience of previous research, this study introduces a larger dataset containing 300 commonly used ASL sentences. Without deliberate selection, these sentences are composed of 455 different hand gestures. Two kinds of labels under the grammar of both sign language and spoken language are added to each sentence. Thus, two NLP models for both sequence recognition and end-to-end translation are also introduced.
The rest of the paper is organized as follows: Section 2 introduces the dataset collection process and the structure of deep learning models. After training the models with the collected data, Section 3 presents the results and evaluation of the models’ outputs. Then, the user-independent validation and the rationality of the methodology are discussed in Section 4. Finally, the conclusion of the paper is drawn.

2. Materials and Methods

The grammar of sign language can be different from the spoken language. For example, the expression “What’s your name” in ASL is the signs’ sequence of “your”, “name”, and “what”. It is also critical to emphasize the subject in declarative or interrogative sentences. When expressing “I am happy”, people tend to sign “I”, “happy”, and “I” to emphasize the protagonist of the topic at the end of the sentence. In this research, we prepare two kinds of labels for one sign language expression. One is under the order of hand gestures, and another is under the rule of spoken language.
Except for grammar, the range of hand movements varies from person to person. In Figure 1a, the expressions of “Sunday” differ in the motion trajectories. Also, people can hardly keep the same movements for one specific word appearing in different sentences. Because the other words around that specific word are not the same, the gesture’s starting state (the ending state of the last word) can be different. One sign language word may also have similar but not identical expressions. As shown in Figure 1b, both gestures are for the word “you”, but they are slightly different in angle.
To solve the problem mentioned above, in this section, the motion capture system is utilized to collect the inertial data of arms and hands during sign language performance. Two deep learning models are introduced for sign language recognition and end-to-end sign language translation.

2.1. Dataset Collection

The upper-body movement is captured by the Perception Neuron (Noitom Ltd., Beijing, China) motion capture system. It includes 25 inertial measurement units (IMUs) fixed by fabric and straps spreading around the back, head, arms, and hands. The distribution of the IMU sensors is shown in Figure 2. Each sensor consists of an accelerometer, gyroscope, and magnetometer and has a small mass and size. Therefore, the device maintains a very light mass overall, even with the use of many sensors. Moreover, comprehensive human movement information is captured by the device and the data quality is high. Without the use of metal parts, the device doesn’t impede the body’s movement and is well-suited for wearing.
The device transfers data to the computer via USB or Wifi. Basic settings and sensor calibration are performed on the supporting software before use. The length of bones between the joints of the human body is considered to be fixed by selecting the user’s height at first. The sampling rate is 60 Hz. At each sampling moment, the data returned by the device is the calculated rotation data (Rotation_Y, Rotation_X, Rotation_Z) of each upper-body joint. The hip joint is a reference point recording the absolute coordinates under the earth coordinate system. The coordinates of other joints are relative coordinates based on their respective reference joints. Figure 3a shows the reference relationship between the coordinates (except for the hands) when performing the initial sensor calibration. The body is in one plane with both arms stretched horizontally. Taking the right arm as an example, the Right Arm takes the Right Shoulder as the reference, and the coordinate values indicate the difference in rotation angle around the three axes between the two. The same relationship exists between other joints. The coordinates of the Right Arm, Right Forearm, and Right Hand in this state are 0. In the next frame, the right arm is slightly raised leading to the positional change in the XOY plane, as shown in Figure 3b. The three mentioned joints produce a rotation around the Z-axis. The coordinate values of Rotation_Z become α 1 , α 2 , and α 3 in the current frame. The relative information of the hand joints is added in Figure 4.
The device captures the rotation data of 59 upper-body joints, but only a part of them are helpful. Due to the limitation of the device, some finger joint movements only have data in the direction of extension/flexion. Moreover, some coordinates of joints (like hip, head, neck, and spine) keep the same value throughout the experiment because of the standing still status during the sign language performance. We manually remove these useless data, and the remaining coordinates are listed in Table 2. Finally, a 38-dimensional vector is selected to describe the state of arms and hands at each moment.
Through the survey of sign language online courses, 300 ASL sentences (some examples are listed in Table A1), including the basic topics of everyday life (weather, daily routine, age, etc.) are selected as the target sentences. Three ASL beginners (female, age: 23~29, height: 158~163 cm, weight: 43~50 kg) joined the experiment after taking the online courses. The data collection process lasted a month. Participant 1 contributed the data for all 300 sentences. Participants 2 and 3 contributed the data for 50 and 20 sentences, respectively. Each sentence was required to be repeated 20 times. Finally, the data amount of the collected ASL dataset was 7400.
Each sentence of sign language lasts for around 1~7 s. Under the sampling rate of 60 Hz, the data along the time direction contains 60~420 points. In the preprocessing of collected inertial data, a median filter is applied to make the data smooth. Then, the data is segmented into frames using the sliding window method. The window size is 36, and the sliding size is 18.

2.2. Model Structure

2.2.1. Sign Language Recognition Model

A sign language sentence is made up of a series of hand gestures. In the sign language recognition task, the label is a series of words sharing the same order with hand gestures. So, hand gestures are required to be recognized correctly in order. In the sequence recognition model of Figure 5, the first layer is the CNN that extracts features from each input data frame. Based on shared weights architecture, CNN eliminates effects from motion differences in amplitude and trajectory [28]. The second layer is Bi-directional Long-Short Term Memory (Bi-LSTM). LSTM can preserve long-term dependencies by controlling the percentage of previous information dropping, current information inputting, and current information outputting [29]. In addition, Bi-LSTM utilizes both forward and backward information on each time step to better contextualize the output. By using the sliding window method in data preprocessing, the output sequence from the Bi-LSTM layer is much longer than the number of words in the label. Therefore, a CTC layer is added to realize the alignment between the output predictions and the label [30]. When using CTC as the loss to train the model, it calculates the sum of probabilities of all possible alignments.
l o s s = l o g p ( a l i g n m e n t | i n p u t ) .
Although not only one alignment leads to the correct result, the output with the largest probability of each frame is chosen as the final result when doing the model validation.
In the preprocessing of text labels, a vocabulary with 455 words (listed in Table A2) for the 300 sentences is built. Thus, by looking up the vocabulary, a word can be represented by the index. Finally, a sentence is converted into a string of numbers in the range of 0~454.

2.2.2. End-to-End Sign Language Translation Model

In the end-to-end sign language translation task, the label sentence has the same grammar rule as the spoken language in daily life. However, some words like “am”, “is”, and “to” do not have corresponding expressions in sign language. Inspired by neural machine translation, an encoder-decoder structured model with global attention is applied to realize the end-to-end sign language translation. As shown in Figure 6, the encoder keeps nearly the same structure as the model in Section 2.2.1. The decoder consisting of the LSTM cell continues to output decoded words until the “end” marker is reached. Global attention between the encoder and decoder is added to learn the mutual mapping relationship between inertial data and the text sentence [31]. In the training step of the model, the input to the decoder in each time step is the combination of the real label word and the result of the last time step. The output of the current time step is determined by the output of the LSTM cell and the attention result from the encoder. The loss of the decoder is the cross entropy between the output distribution and the label word. The CTC loss from the encoder is also considered together to optimize the model.
The translation task’s vocabulary differs from the above-mentioned sequence recognition, so another vocabulary (listed in Table A3) is built for the label sentences under the spoken language grammar. In addition to the 502 ordinary words, two special tokens “<BOS>” and “<EOS>” representing “beginning of sentence” and “end of sentence” are also added to the vocabulary. All 300 sentences have these two special tokens added at the beginning and the end.

3. Results

3.1. Sign Language Recognition Results

The collected 7400 sentences are randomly divided into five equal parts. To do five-fold cross-validation, one part is taken as the validation set each time, and the other four parts are collectively used as the training set. The training set contains 5920 sentences and the validation set contains 1480 sentences. The deep learning model is built with Python 3.7 (Python Software Foundation, Beaverton, OR, USA) and PyTorch 1.13.1 (Meta AI, New York City, NY, USA). In the training step, the optimizer is Adam, with a learning rate of 0.0001. After training the model for 20 epochs, the CTC loss drops to a low level. The recognition results of the model on the validation set are listed in Table 3. The word error rate (WER) measures the least operations of substitution, deletion, and insertion to transform the predicted sequence into the label sequence.
W E R = N s u b + N d e l + N i n s N a l l   w o r d s ,
where Nsub, Ndel, and Nins are the numbers of necessary substitutions, deletions, and insertions, respectively. The sentence error rate (SER) measures the percentage of not entirely correct sequences of the validation results.
S E R = N e r r o r   s e q u e n c e s N a l l   s e q u e n c e s
When treating the sign language sentence as a sequence of hand gestures, the model can recognize each sequence with a high accuracy rate. In addition, the alignments between the model output and the label words are also well achieved. This shows that the data collected by the device can accurately represent different gestures.

3.2. End-to-End Sign Language Translation Results

The end-to-end translation model is trained with the same settings as the above-mentioned model. The error rates for both the word and sentence-level evaluation increase considerably in the end-to-end translation results, as illustrated in Table 4. Since the sentences outputted by the decoder tend to reach the “<EOS>” token earlier, the output sentences from the model are generally shorter than the label sentences, leading to a large number of insertion errors. Spoken language words that could not be expressed by hand gestures also show higher error rates in both substitution and insertion. Due to the small number of text sentences, the model could learn limited knowledge about English grammar, providing a limited enhancement effect on the translation results.
The end-to-end translation model utilizes both CTC loss and cross-entropy loss, and the labels for each type of loss come from different vocabularies. To verify the role of CTC loss, validation fold 4 which is closest to the average result is selected as the validation set. Then, the model is retrained after removing the CTC loss. From the results of the validation set, both WER and SER increase by 1.49% and 0.27%, respectively. This indicates that CTC loss promotes model optimization.
The recognition and translation have two different vocabularies because the words that make up the two kinds of labels are not the same. When the two vocabularies are unified into one vocabulary, word indexes are also modified. When the model is trained with the modified labels, the resulting error rates are also increased by 0.97% and 3.18% compared to the original results, as shown in Figure 7. Using only one vocabulary causes the labels of CTC loss to be sparse, and many words in the vocabulary do not appear on CTC labels. This makes alignment more difficult, leading to an increase in error rates.

4. Discussion

4.1. User-Independent Validation

4.1.1. Word-Level User-Independent Validation

Participants 2 and 3 only attended a part of the experiment (contributing 50 sentences and 20 sentences, respectively), and the sentences they contributed were quite different from each other. So, 20 words (listed in Table 5) for which all three participants have data recorded are selected for word-level user-independent validation. The dataset of these words is manually segmented from the sentences. The same word may come from different sentences, so the starting and ending states of the gestures for the same word could be different.
The classifier consists of the CNN feature extractor and a fully connected classification layer. The word data from one participant is used as the validation set, and the remaining data from the other two participants is used to train the classifier. The classification result is shown in Figure 8a. The average classification accuracy of all participants keeps a high level of around 88%. This illustrates that the gestures for the same word have distinguishable common features even from different people or different sentences. The total confusion matrix of these 20 words from 3 participants is shown in Figure 8b. The word “finish” is easy to be recognized as the word “sad”, and the word “you” is also confused with “I” and “we”. This is because these two groups of words have similar hand shapes, but the hand movements are different.

4.1.2. Sentence-Level User-Independent Validation

Participant 2 contributed 1000 pieces of data is regarded as the testing set in the sentence-level user-independent validation. The sign language recognition model introduced in Section 2.2.1 is applied here to only recognize a sequence of gestures without considering the grammar rule. The model is trained with the 6400 pieces of data from the remaining two participants. The testing results of Participant 2 are listed in Table 6. The substitution errors increase a lot, indicating that the recognition ability of the model drops dramatically under the condition of fewer participants. Due to the low number of experimental participants, the model could not learn enough patterns from the training set.

4.2. Sequence Recognition with Encoder-Decoder Model

In the sign language recognition task of Section 2.2.1, the model recognized hand gestures for each input data frame and did the alignment between the output frames and the label words with the CTC layer. Since it is a sequence-to-sequence task, it can also be achieved by an encoder-decoder structured model. Using the same model as Section 2.2.2, the decoder can generate the sequence recognition results by time steps. The recognition results with the encoder-decoder model are shown in Table 7. All three errors increase significantly compared with the CTC-based sequence recognition model. The output from the model tends to reach the “<EOS>” earlier, leading to a large number of insertion errors. Also, the decoder could not recognize gestures correctly using the information from the encoder, causing higher substitution errors.

4.3. Advantages and Limitations

Compared with vision-based methods, wearable sensors can obtain more direct human movement data. Recognition of sign language videos can be affected by illumination, background environment, and occlusions, but this problem will not happen while using the sensors’ data. Moreover, the real-life application of the camera involves the security of personal information and will inevitably record information about other people in the surrounding area. Most currently existing vision-based datasets are RGB videos without considering depth information. In contrast, the motion capture system restores the movement of the human body in three-dimensional space.
The proposed method has certain limitations and spaces for improvement. In this research, the movements of 300 ASL sentences are captured by the motion capture system and recognized by the deep learning model. However, it is still far away from being applied in everyday life. In addition to the limited number of sign language sentences, more external factors, such as the environment, need to be considered. The sequence recognition model can recognize each gesture with high accuracy. However, the accuracy drops in the end-to-end translation due to the difference in grammar between sign language and spoken language. A language model is needed to make the output sentences closer to spoken language. In sentence-level user-independent validation, the accuracy dropped dramatically due to the inter-individual differences in movement. More participants should be involved to let the model learn more patterns from a wide variety of data. The off-the-shelf motion capture system contains multiple inertial sensors. With the development of wearable sensor technology, more convenient devices could be considered for the research on sign language translation.

5. Conclusions

This research presented the basic process of sign language recognition and end-to-end translation by wearable sensors. An ASL dataset with 300 sentences for daily use was collected by the inertial motion capture system. Two kinds of deep learning models were constructed based on the grammar rule of both sign language and spoken language. Generally, the sequence recognition model achieved relatively high accuracy without considering individual differences. In contrast, the end-to-end translation presented more errors due to the lack of grammar knowledge. In the user-independent validation, the selected dataset with limited words showed high classification accuracy. In the sentence-level validation, the vocabulary containing more words increased the recognition difficulty, and the accuracy rate decreased.

Author Contributions

Conceptualization, Y.G. and M.T.; methodology, Y.G. and M.T.; software, Y.G.; validation, Y.G.; formal analysis, Y.G.; investigation, Y.G.; data curation, Y.G.; writing—original draft preparation, Y.G.; writing—review and editing, H.O.; supervision, H.O.; project administration, M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by Gunma University for the promotion of scientific research.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the participants provided their written informed consent to participate in this study.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Examples of sentences under both sign language and spoken language grammar.
Table A1. Examples of sentences under both sign language and spoken language grammar.
Sign Language SentencesSpoken Language Sentences
nice meet younice to meet you
i learn sign iI learn sign
your name whatwhat’s your name
you from wherewhere are you from
your favorite movie whatwhat’s your favorite movie
you like what_dowhat do you like to do
you work wherewhere do you work
time whatwhat time
bathroom wherewhere is bathroom
please sign again slowerplease sign again slower
i have two dog one catI have 2 dogs and 1 cat
my favorite sport basketballmy favorite sport is basketball
i want pancake sausageI want pancakes and sausages
meet new people my son become shywhen meeting new people my son becomes shy
you look worry youyou look worried
lion wherewhere is the lion
elephant hugeelephant is huge
my book wherewhere is my book
you like icecream youdo you like ice-cream
more soup pleasemore soup please
you want fries alsodo you also want fries
wow steak deliciouswow the steak is delicious
happy new yearhappy new year
christmas you want what youwhat do you want for Christmas
tomorrow weather whatwhat’s the weather tomorrow
senior year i take calculus iI take calculus in senior year
picnic i bring hotdog iI bring hotdogs for picnic
i think future maybe become biologistI think I may become a biologist in future
laughing commercial funnyhaha the commercial is funny
sunday you want watch super_bowl youdo you want to watch super bowl on Sunday
Table A2. The sign language recognition vocabulary.
Table A2. The sign language recognition vocabulary.
AC, AUG, US, a_little, a_lot, across_street, afraid, afternoon, again, age, alcohol, allergic, alright, also, altogether, always, angry, animal, annoy, apartment, apple, area, arrive, baby, back, bad_at, bake, baseball, basketball, bathroom, beach, because, become, before, biologist, birthday, black, blonde, blood, book, borrow_me, boxing, bring, broke, brother, brown, brush_teeth, busy, but, cake, calculus, canada, car, cash, cat, celebrate, chair, chemistry, chocolate, christmas, christmas_day, church, class, cleaning, clinic, close, clothes, cloud, coat, coffee, cold, college, color, come, come_on, comfortable, commercial, compete, congratulations, cook, cookies, cost, costume, country, cousin, credit, cute, dad, dark, daughter, day, deaf, delicious, department, dessert, different, dinner, dislike, do_you, dog, dollars, drink, drug, dry, due, during, easter, eat, eat_breakfast, egg, eight, elephant, engineering, english, enjoy, every_day, every_monday, every_morning, every_saturday, every_thursday, excited, excuse_me, experience, fall, family, fancy, favorite, feel, fifteen, fine, finger_spell, finish, five, five_oclock, flower, food, football, for, forget, friday, friend, fries, from, funny, future, game, get_together, ghost, give_me, go, go_ahead, go_out, good, good_at, goodbye, graduate, grandma, growing_up, hail, hair, half_time, halloween, hamburger, happy, hat, hate, have, hawaii, he, hearing, hearing_aid, heart, hello, help, help_you, here, hey, high, hiking, history, holiday, home, hope, host, hot, hotdog, hotel, house, how, how_many, how_old, huge, hungry, hunt, husband, i, icecream, important, introduce, jingle_bells, join, karate, ketchup, knit, know, lake, language, large, last, latter, laughing, learn, leave, lets, light, like, lion, live, local, look, look_around, lost, love, love_all, love_it, lunch, machine, major, make, manage, many, marry, maybe, medicine, medium, meet, meeting, milk, mix, monday, money, monkey, more, morning, mother, mountain, move, movie, museum, mustache, my, name, nauseous, need, nervous, new, next_to, next_week, nice, night, nine, nine_oclock, nineteen, no, not, not_know, not_yet, now, office, olympics, one, optometrist, orange, order, other, our, pack, pain, pancake, paper, parade, party, pass_out, pay, people, physics, picnic, pineapple, pink, plan, play, please, police_officer, potato, prayer, pregnant, president, pressure, principal, problem, pumpkin, purse, rabbit, rain, rainbow, read, ready, red, reduce, remember, resolution, rest, restaurant, run, run_out, sad, same, saturday, sausage, scared, scarf, school, season, seatbelt, see, seizures, senior, seven, seven_dollars, she, shirt, shop, show, shrimp, shy, sign, since, sister, six_oclock, skill, slower, small, snow, soda, sometimes, son, soup, spaghetti, spider, sport, spread, spring, start, state, stay, steak, still, strawberry, strong, student, study, subscribe, summer, sunday, sunny, super_bowl, support, sweet, swimming, take, take_care, take_out, talkative, teacher, team, tend, test, thank_you, thankful, thanksgiving, thanksgiving_day, that, they, think, thirty, this, three, three_oclock, three_of_you, thursday, time, tomorrow, tornado, trash, travel, treat, trick, try, turkey, turn_on, twelve_oclock, twenty_six, two, two_of_us, two_week, uncle, understand, use, vacation, valentine, vegetable, visit, vote, waitress, walk, want, warm, wash, watch, we, wear, weather, weekend, welcome, what, what_do, what_time, what_year, whats_up, when, where, which, white, who, whole, why, wife, win, winter, witch, with, work, working, workout, worry, wow, wrestling, write, year, yes, you, you_all, you_mind, your, yourself.
Table A3. The sign language translation vocabulary.
Table A3. The sign language translation vocabulary.
1, 12, 15, 19, 2, 26, 3, 30, 31, 35, 5, 6, 7, 87, 9, :, AC, August, Canada, Christmas, Easter, English, Friday, Halloween, Hawaii, I, Monday, Saturday, Sunday, Thanksgiving, Thursday, US, Valentine’s, a, across, afraid, afternoon, again, age, ahead, alcohol, all, allergic, alright, also, altogether, always, am, an, and, angry, animal, annoyed, apartment, apple, are, area, arrive, as, at, baby, back, bad, bake, baseball, basketball, bathroom, be, beach, because, become, becomes, before, bells, biologist, birthday, birthdays, black, blonde, blood, book, borrow, both, bowl, boxing, breakfast, bring, bringing, broke, brother, brown, brush, busy, but, cakes, calculus, car, care, cash, cat, celebrating, chair, chemistry, chocolate, church, class, cleaning, clinic, close, clothes, clouds, cloudy, coat, coffee, cold, college, color, come, comfortable, coming, commercial, competing, congratulations, cooked, cookies, cooking, cost, costs, costume, country, cousin, credit, cute, dad, dark, daughter, day, deaf, delicious, department, dessert, did, different, dinner, dislike, dislikes, do, does, dog, dogs, dollars, drinking, drugs, dry, due, during, eat, eating, eggs, elephant, engineering, enjoy, enjoys, every, everyday, excited, excuse, experience, fall, family, fancy, favorite, feel, fine, finger, finish, finished, flowers, food, football, for, forget, friend, friends, fries, from, funny, future, game, get, ghost, give, go, going, good, goodbye, graduate, graduation, grandma, growing, had, haha, hail, hair, half, hamburgers, happy, has, hat, hate, have, hearing, hearing-aid, heart, hello, help, here, hey, high, hiking, him, his, history, holidays, home, hope, host, hot, hotdogs, hotel, house, how, huge, hungry, hunting, husband, ice-cream, important, in, introduce, is, it, jingle, join, karate, ketchup, knit, know, lake, language, large, last, latter, learn, leave, let’s, light, like, lion, little, live, lives, local, look, lost, lot, love, lunch, machine, major, make, manage, many, marry, may, me, medicine, medium, meet, meeting, merry, milk, mind, mix, mom, money, monkey, more, morning, mother, mother’s, mountain, moved, movie, movies, much, museum, mustache, my, name, nauseous, need, needed, nervous, new, next, nice, night, no, not, now, of, office, officer, old, olympics, on, optometrist, or, orange, order, other, others, our, out, packing, pain, pancakes, paper, parade, party, pass, pay, people, physics, picnic, pineapple, pink, plan, play, please, police, potato, prayer, prefer, pregnant, president, pressure, principal, problem, pumpkin, purse, rabbits, rain, rainbow, rains, rainy, read, ready, red, reduction, remember, resolution, restaurant, resting, running, sad, same, sausages, scared, scarf, school, season, seatbelt, see, seizures, senior, shirt, shop, show, shrimps, shy, sign, since, sister, skills, slower, small, snow, soda, sometimes, son, soup, spaghetti, spell, spider, sport, spread, spring, start, starts, state, states, stay, steak, still, strawberry, street, strongly, student, studying, subscribe, summer, sunny, super, support, sweet, swimming, take, talkative, teacher, team, teeth, tends, test, thank, thankful, that, the, there, they, think, this, three, time, to, today, together, tomorrow, too, tornado, trash, travel, treat, trick, try, turkey, turning, two, uncle’s, understand, up, use, vacation, vegetable, visit, vote, waitress, walking, want, wants, warm, wash, watch, watching, we, wear, weather, week, weekends, weeks, welcome, what, what’s, when, where, which, white, who, whole, whose, why, wife, will, win, winter, witch, with, work, working, workout, worried, would, wow, wrestling, write, year, yes, yet, you, your, yourself, <BOS>, <EOS>.

References

  1. World Health Organization. Available online: https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss (accessed on 25 April 2023).
  2. Sun, C.; Zhang, T.; Xu, C. Latent support vector machine modeling for sign language recognition with Kinect. ACM Trans. Intell. Syst. Technol. 2015, 6, 20. [Google Scholar] [CrossRef]
  3. Fang, B.; Co, J.; Zhang, M. DeepASL: Enabling Ubiquitous and Non-Intrusive Word and Sentence-Level Sign Language Translation. In Proceedings of the 15th ACM Conference on Embedded Network Sensor Systems, Delft, The Netherlands, 6–8 November 2017. [Google Scholar] [CrossRef]
  4. Koller, O.; Forster, J.; Ney, H. Continuous sign language recognition: Towards large vocabulary statistical recognition systems handling multiple signers. Comput. Vis. Image Underst. 2015, 141, 108–125. [Google Scholar] [CrossRef]
  5. Zhou, H.; Zhou, W.; Li, H. Dynamic Pseudo Label Decoding for Continuous Sign Language Recognition. In Proceedings of the IEEE International Conference on Multimedia and Expo, Shanghai, China, 8–12 July 2019. [Google Scholar] [CrossRef]
  6. Zhang, Q.; Wang, D.; Zhao, R.; Yu, Y. MyoSign: Enabling End-to-End Sign Language Recognition with Wearables. In Proceedings of the 24th International Conference on Intelligent User Interfaces, New York, NY, USA, 17–20 March 2019. [Google Scholar] [CrossRef]
  7. Hou, J.; Li, X.Y.; Zhu, P.; Wang, Z.; Wang, Y.; Qian, J.; Yang, P. Signspeaker: A Real-Time, High-Precision Smartwatch-Based Sign Language Translator. In Proceedings of the 25th Annual International Conference on Mobile Computing and Networking, Los Cabos, Mexico, 21–25 October 2019. [Google Scholar] [CrossRef]
  8. Zhang, Q.; Jing, J.; Wang, D.; Zhao, R. Wearsign: Pushing the Limit of Sign Language Translation Using Inertial and EMG Wearables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022, 6, 35. [Google Scholar] [CrossRef]
  9. Wang, F.; Zhao, S.; Zhou, X.; Li, C.; Li, M.; Zeng, Z. An recognition–verification mechanism for real-time Chinese sign language recognition based on multi-information fusion. Sensors 2019, 19, 2495. [Google Scholar] [CrossRef] [PubMed]
  10. Paudyal, P.; Lee, J.; Banerjee, A.; Gupta, S.K. A comparison of techniques for sign language alphabet recognition using armband wearables. ACM Trans. Interact. Intell. Syst. 2019, 9, 14. [Google Scholar] [CrossRef]
  11. Tateno, S.; Liu, H.; Ou, J. Development of sign language motion recognition system for hearing-impaired people using electromyography signal. Sensors 2020, 20, 5807. [Google Scholar] [CrossRef] [PubMed]
  12. Kim, S.; Kim, J.; Ahn, S.; Kim, Y. Finger language recognition based on ensemble artificial neural network learning using armband EMG sensors. Technol. Health Care 2018, 26, 249–258. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, Z.; Zhao, T.; Ma, J.; Chen, H.; Liu, K.; Shao, H.; Wang, Q.; Ren, J. Hear sign language: A real-time end-to-end sign language recognition system. IEEE Trans. Mob. Comput. 2020, 21, 2398–2410. [Google Scholar] [CrossRef]
  14. Savur, C.; Sahin, F. American Sign Language Recognition System by Using Surface EMG Signal. In Proceedings of the 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Budapest, Hungary, 9–12 October 2016. [Google Scholar] [CrossRef]
  15. Jane, S.P.; Sasidhar, S. Sign Language Interpreter: Classification of Forearm EMG and IMU Signals for Signing Exact English. In Proceedings of the 2018 IEEE 14th International Conference on Control and Automation (ICCA), Anchorage, AK, USA, 12–15 June 2018. [Google Scholar] [CrossRef]
  16. Paudyal, P.; Banerjee, A.; Gupta, S.K. Sceptre: A Pervasive, Non-Invasive, and Programmable Gesture Recognition Technology. In Proceedings of the 21st International Conference on Intelligent User Interfaces, Sonoma, CA, USA, 7–10 March 2016. [Google Scholar] [CrossRef]
  17. Goel, S.; Kumar, M. A Real Time Sign Language Interpretation of Forearm Based on Data Acquisition Method. In Proceedings of the 2019 International Conference on Signal Processing and Communication, Noida, India, 7–8 March 2019. [Google Scholar] [CrossRef]
  18. Savur, C.; Sahin, F. Real-Time American Sign Language Recognition System Using Surface emg Signal. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications, Miami, FL, USA, 9–11 December 2015. [Google Scholar] [CrossRef]
  19. Yu, Y.; Chen, X.; Cao, S.; Zhang, X.; Chen, X. Exploration of Chinese sign language recognition using wearable sensors based on deep belief net. IEEE J. Biomed. Health Inform. 2019, 24, 1310–1320. [Google Scholar] [CrossRef] [PubMed]
  20. Li, Y.; Chen, X.; Tian, J.; Zhang, X.; Wang, K.; Yang, J. Automatic Recognition of Sign Language Subwords Based on Portable Accelerometer and EMG Sensors. In Proceedings of the International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction, Beijing, China, 8–12 November 2010. [Google Scholar] [CrossRef]
  21. Ambar, R.; Fai, C.K.; Abd Wahab, M.H.; Jamil, M.M.; Ma’radzi, A.A. Development of a wearable device for sign language recognition. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2018; Volume 1019, p. 012017. [Google Scholar] [CrossRef]
  22. Lee, B.G.; Chong, T.W.; Chung, W.Y. Sensor fusion of motion-based sign language interpretation with deep learning. Sensors 2020, 20, 6256. [Google Scholar] [CrossRef] [PubMed]
  23. Wen, F.; Zhang, Z.; He, T.; Lee, C. AI enabled sign language recognition and VR space bidirectional communication using triboelectric smart glove. Nat. Commun. 2021, 12, 5378. [Google Scholar] [CrossRef] [PubMed]
  24. Korzeniewska, E.; Kania, M.; Zawiślak, R. Textronic Glove Translating Polish Sign Language. Sensors 2022, 22, 6788. [Google Scholar] [CrossRef]
  25. Liu, Y.; Jiang, F.; Gowda, M. Finger Gesture Tracking for Interactive Applications: A Pilot Study with Sign Languages. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 112. [Google Scholar] [CrossRef]
  26. Jin, Y.; Gao, Y.; Zhu, Y.; Wang, W.; Li, J.; Choi, S.; Li, Z.; Chauhan, J.; Dey, A.K.; Jin, Z. SonicASL: An Acoustic-based Sign Language Gesture Recognizer Using Earphones. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 67. [Google Scholar] [CrossRef]
  27. Perception Neuron Products. Available online: https://neuronmocap.com/perception-neuron-series (accessed on 25 April 2023).
  28. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. In The Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, MA, USA, 1995; Volume 3361. [Google Scholar]
  29. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks. In Proceedings of the 23rd International Conference on Machine Learning, New York, NY, USA, 25–29 June 2006. [Google Scholar] [CrossRef]
  31. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
Figure 1. Different expressions for the same word: (a) “Sunday”; (b) “you”.
Figure 1. Different expressions for the same word: (a) “Sunday”; (b) “you”.
Sensors 24 00453 g001
Figure 2. Sensor distribution under the human back view [27].
Figure 2. Sensor distribution under the human back view [27].
Sensors 24 00453 g002
Figure 3. The relative relationship of upper-body joint coordinates: (a) initial sensor calibration posture; (b) right arm slightly raised.
Figure 3. The relative relationship of upper-body joint coordinates: (a) initial sensor calibration posture; (b) right arm slightly raised.
Sensors 24 00453 g003
Figure 4. The hierarchy of all joint coordinate data returned by the device.
Figure 4. The hierarchy of all joint coordinate data returned by the device.
Sensors 24 00453 g004
Figure 5. The sequence recognition model for the sign language recognition task.
Figure 5. The sequence recognition model for the sign language recognition task.
Sensors 24 00453 g005
Figure 6. The encoder-decoder structured model for the sign language translation task.
Figure 6. The encoder-decoder structured model for the sign language translation task.
Sensors 24 00453 g006
Figure 7. Comparative results of translation under different conditions.
Figure 7. Comparative results of translation under different conditions.
Sensors 24 00453 g007
Figure 8. Word-level user-independent validation results: (a) Classification accuracy of three participants; (b) Confusion Matrix of all 20 words.
Figure 8. Word-level user-independent validation results: (a) Classification accuracy of three participants; (b) Confusion Matrix of all 20 words.
Sensors 24 00453 g008
Table 1. The summary of previous studies on sign language recognition.
Table 1. The summary of previous studies on sign language recognition.
ReferenceWearablesNo. of WordsNo. of SentencesNo. of Participants
[6]Myo armband7010015
[7]Smartwatch1037315
[8]Myo armband & Smartwatch10025015
[14]Myo armband26-10
[15]Myo armband48-1
[16]Myo armband30-10
[19]EMG & IMU sensors150-8
[20]EMG & ACC sensors121-1
[24]Customized data glove26-3
[25]Smartwatch & Smart ring100-10
[26]Speaker & Microphone42308
Table 2. The selected coordinates for sign language recognition.
Table 2. The selected coordinates for sign language recognition.
JointsCoordinates
Right & Left ArmRotation_Y, Rotation_X, Rotation_Z
Right & Left ForearmRotation_Y, Rotation_X, Rotation_Z
Right & Left HandRotation_Y, Rotation_X, Rotation_Z
Right & Left Hand Thumb 1Rotation_Y, Rotation_Z
Right & Left Hand Thumb 2Rotation_Y, Rotation_Z
Right & Left Hand Thumb 3Rotation_Y
Right & Left Hand Index 1Rotation_Z
Right & Left Hand Index 2Rotation_Z
Right & Left Hand Middle 1Rotation_Z
Right & Left Hand Ring 1Rotation_Z
Right & Left Hand Pinky 1Rotation_Z
Table 3. The sign language recognition results.
Table 3. The sign language recognition results.
FoldNo. of WordsNo. of SentencesInsertion
Errors
Deletion
Errors
Substitution
Errors
WERSER
155521480113270.74%2.09%
25616148061010.30%1.15%
355661480636351.38%3.65%
455851480834351.38%3.85%
5562114801514180.84%2.57%
Average55881480721230.93%2.66%
Table 4. The end-to-end sign language translation results.
Table 4. The end-to-end sign language translation results.
FoldNo. of WordsNo. of SentencesInsertion
Errors
Deletion
Errors
Substitution
Errors
WERSER
16925148055111426313.40%34.19%
2696014809919747022.39%47.03%
3691314806656720713.58%30.95%
46964148054620843817.12%39.93%
56938148050121344316.68%44.26%
Average6940148065114036416.63%39.27%
Table 5. Word-level user-independent validation with 20 ASL words.
Table 5. Word-level user-independent validation with 20 ASL words.
Word Index0123456789
Selected Wordalsochurcheatfeelfinishfriendfriesgo tohappyhave
Word Index10111213141516171819
Selected WordImysadsummerSundaythisweyearyoudislike
Table 6. Sentence-level user-independent validation results.
Table 6. Sentence-level user-independent validation results.
Substitution
Errors
Deletion
Errors
Insertion
Errors
Total
Errors
No. of Words or SentencesError
Rate
Accuracy
Rate
Word-Level Evaluation1451360711882340055.36%44.64%
Sentence-Level Evaluation---729100072.90%27.10%
Table 7. Sequence recognition result with encoder-decoder model.
Table 7. Sequence recognition result with encoder-decoder model.
Substitution
Errors
Deletion
Errors
Insertion
Errors
Total
Errors
No. of Sentences or WordsError
Rate
Accuracy
Rate
Word-Level Evaluation35852131148055829.69%90.31%
Sentence-Level Evaluation---388148026.22%73.78%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gu, Y.; Oku, H.; Todoh, M. American Sign Language Recognition and Translation Using Perception Neuron Wearable Inertial Motion Capture System. Sensors 2024, 24, 453. https://doi.org/10.3390/s24020453

AMA Style

Gu Y, Oku H, Todoh M. American Sign Language Recognition and Translation Using Perception Neuron Wearable Inertial Motion Capture System. Sensors. 2024; 24(2):453. https://doi.org/10.3390/s24020453

Chicago/Turabian Style

Gu, Yutong, Hiromasa Oku, and Masahiro Todoh. 2024. "American Sign Language Recognition and Translation Using Perception Neuron Wearable Inertial Motion Capture System" Sensors 24, no. 2: 453. https://doi.org/10.3390/s24020453

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop