Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Convolutional and recurrent neural network for human activity recognition: Application on American sign language

  • Vincent Hernandez ,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    vincent.hernandez1985@gmail.com

    Affiliation GVLAB - University of Agriculture and Technology of Tokyo, Tokyo, Japan

  • Tomoya Suzuki,

    Roles Data curation, Software

    Affiliation GVLAB - University of Agriculture and Technology of Tokyo, Tokyo, Japan

  • Gentiane Venture

    Roles Conceptualization, Funding acquisition, Project administration, Resources, Supervision

    Affiliation GVLAB - University of Agriculture and Technology of Tokyo, Tokyo, Japan

Abstract

Human activity recognition is an important and difficult topic to study because of the important variability between tasks repeated several times by a subject and between subjects. This work is motivated by providing time-series signal classification and a robust validation and test approaches. This study proposes to classify 60 signs from the American Sign Language based on data provided by the LeapMotion sensor by using different conventional machine learning and deep learning models including a model called DeepConvLSTM that integrates convolutional and recurrent layers with Long-Short Term Memory cells. A kinematic model of the right and left forearm/hand/fingers/thumb is proposed as well as the use of a simple data augmentation technique to improve the generalization of neural networks. DeepConvLSTM and convolutional neural network demonstrated the highest accuracy compared to other models with 91.1 (3.8) and 89.3 (4.0) % respectively compared to the recurrent neural network or multi-layer perceptron. Integrating convolutional layers in a deep learning model seems to be an appropriate solution for sign language recognition with depth sensors data.

Introduction

Sign language is a language that mainly uses hand kinematics and facial expressions. It is widely used by hearing-impaired people to communicate with each other, but rarely with people who do not have hearing impairment. Therefore, they are only in direct contact with hearing-impaired persons, which greatly limits social interactions. An alternative would be to have a real-time translation with interpreters, but they are not permanently available and can be rather expensive. Therefore, a system that could enable automatic translation would be of great interest.

Human Activity Recognition (HAR) in general is an important and challenging topic o address because of the large variability that exists for a given task. Indeed, whether the variability comes from a subject repeating an action several times or more importantly between subjects, the kinematics behavior over time presents a certain challenge to generalize. HAR considers that these behaviors are represented by specific patterns that could be classified using machine learning algorithms.

Nowadays, human movement data can be easily extracted from low-cost systems that integrate depth mapping sensors such as Kinect and LeapMotion. These systems are ready to use, require a relatively short set-up time and the data can be easily extracted. Therefore, they can be easily used to quickly acquire a large amount of data, which is a requirement when considering machine learning. Other approaches were considered with sEMG [1], CyberGloves [2] or motion capture system but these techniques are difficult to use outside of laboratories.

Considering systems such as the Kinect® or a combined Kinect/LeapMotion® approach, several studies have been conducted using video or depth-mapping sensor and machine learning approaches such as Hidden-Markov-Model (HMM) [3], coupled HMM [4], random-forest per-pixel [5], multi-class Support Vector Machine (SVM) [6], linear binary SVM [7], convolutional neural network (ConvNet) [8,9] and DeepConvLSTM video-based approaches [10] to name a few. Such approaches are interesting, but the Kinect remains difficult to use in public space since it requires a large space and a power supply.

The LeapMotion easily detects forearm and palm movements as well as the position of the fingers and thumb with a measurement accuracy of 200 μm [11] to estimate the joint position. Such accuracy can be useful for creating accurate models of the kinematics of the hand-arm system. In addition, it can be used with a single USB port, consumes very little power, and does not require a large space to be used, making it convenient for applications in off-lab environments. Moreover, it is an affordable system that does not require a professional to set up and no calibration is required, which makes the system practical to use. In addition, the development of depth-sensing cameras on smartphones would be an interesting choice as a basis for sign language recognition and for the development of portable systems that are easy to use.

Naidu and Ghotkar [12] used the LeapMotion to classify a subset of the Indian Sign Language (10 Arabic numbers, 26 letters, and 9 words). They propose four different approaches (Euclidian distance measure, similarity, Jaccard and dice similarity) with 8 distances computed from 6 measured points (center of the palm and all fingertips). Cosine similarity showed an accuracy of 90.00% for the complete dataset but their approach is limited to static posture. Marin et al. [6] studied a combined LeapMotion/Kinect approach with multi-class SVM to classify 10 static American Sign Language (ASL) words with manual extraction of key moments. They presented an average accuracy of 80.86%, 89.71% and 91.28% with the LeapMotion, the Kinect and combined Kinect/LeapMotion approaches respectively with a user-independent k-fold cross-validation (M-1 subject for each K) for parameters tuning. Then, the classifier was trained again with all subjects (M) and computed the accuracy on it which does not provide a reliable model that can be used with new subjects. Kumar et al. [4] also considered a combined Kinect/LeapMotion approach on 25 dynamics words of the Indian Sign Language. They considered a coupled-HMM approach and gained 90.8% accuracy on 25% out of the complete dataset (50% used for training and 25% for parameter tuning and parameter validation), however, they do not provide information whether or not a user-independent test was considered. Chuan et al. [13] classified the 26 letters of the English alphabet using k-NN and SVM classifiers. Results showed an average classification rate of 72.78% and 79.83% respectively using k-fold cross-validation (K = 4) on the complete dataset composed of 2 subjects. Fok et al. [14] proposed an HMM-based approach to recognize the 10 ASL Arabic numbers with an overall average recognition of 93.14% half of the sample of each subject for training and the remaining data to test the model.

The recent breakthrough of deep learning has outperformed conventional machine learning approach in computer vision tasks [15]. Deep learning uses a succession of layers to extract information, with the output of one layer as the input of the following one. They provide a robust approach to generalize inter-subject variability and can consider the time-series dynamics behavior of human movements mainly with ConvNet and Recurrent Neural Network (RNN). Regarding sign language recognition, Koller et al. [16] use a hybrid ConvNet-HMM approach based on a sequence of images extracted from video and by tracking the dominant hand.

Regarding deep learning, the state of the art on HAR is presented by Ordóñez and Roggen [17] adapted from Sainath et al. [18] for speech recognition. They proposed a model based on combining convolutional and recurrent layers with Long-Short-Term-Memory (LSTM) called DeepConvLSTM with data from Inertial Measurement Unit sensors (IMUs). ConvNet and RNN have supervised learning approaches that can learn the dependencies between given inputs and outputs. DeepConvLSTM outperformed ConvNet and RNN considered independently for speech recognition [18] and is considered as the state-of-the-art for HAR [17,19]. Indeed, both approaches have their advantages with convolutional layers that can extract features from a given signal and recurrent layers that can consider the dynamics of a time-series signal [2022]. A combined approach seems to be the best approach to perform HAR and is promising since it allows flexible data fusion and features extraction [17] and this approach will be exploited in this study.

This research proposes a comparison of deep neural networks to classify 60 ASL signs based on a complete kinematic model of the right and left forearm/hands/fingers/thumb. The kinematics of the hand and forearm are derived from the skeletal tracking model provided by the LeapMotion sensor. The purpose of this research is also to propose a robust user-independent k-fold cross-validation and tests in contrast to previous studies that focused primarily on intra-user testing of their models. Indeed, machine learning models can be representative of the behavior of participants but not on new ones. It is also demonstrated that intra-user tests can lead to an overestimation of accuracy that is not representative of the results that can be obtained on new users.

This document is organized as follows. We present the experiment as well as details on the extraction and processing of the data prior to their use in our models. Then, we present the classification methods of the gestures, the learning properties of the models and the hyperparameters used. Finally, the results comparing the different models are presented and discussed.

Material and methods

Experimentation

The experiment was approved by the local ethics committee at the Tokyo University of Agriculture and Technology (TUAT) in Koganei, Japan. All Experiments were done in 2018. Before the experiment, the participants gave informed consent to participate in the study.

A dataset of 25 male subjects, all novices in any sign language, was collected. Before each measurement, the corresponding sign was taught to them. In this study, the Arabic numbers from 0 to 10 and 49 words were considered. The total numbers of each sign gathered are presented in Table 1.

thumbnail
Table 1. Numbers of sign gathered during the experiment with the 25 subjects.

https://doi.org/10.1371/journal.pone.0228869.t001

Features extraction

The LeapMotion SDK Unity core assets 4.3.2 was used with the C#/Unity module to collect data during the experimental protocol. A total number of 26 points on each hand are extracted in real-time from the LeapMotion. The features computed from these points are based on a multi-finger modeling approach presented by Carpinella et al. [23] with minor modifications and the ISB recommendations for the relative orientation of the hand to the forearm [24].

Kinematics models

The following coordinates systems are considered for the hand (1) and forearm (2) (Fig 1). (1) has the direction of (M2-M5) pointing externally, is perpendicular to the plan form by (M2-RS) and (P5-RS) pointing forwardly, and is the cross product of pointing upwardly. (2) has the direction of (EL-US) pointing upwardly, is the cross product of and (RS-US) pointing forwardly and is the cross product of and pointing externally. Then, the Euler angles that describe the relative orientation of the hand with the forearm (flexion/extension and radial/ulnar deviation) are computed with a ZXY rotation sequence (Vectors direction are expressed from the standard anatomical position).

thumbnail
Fig 1. Hand joints.

Thumb, index, middle, ring, and pinky are numbered from 1 to 5 respectively. EFi (i = 1–5): finger and thumb end-effector position, Mi (i = 1–5): finger and thumb metacarpophalangeal joint, Pi (i = 1–5): finger and thumb proximal interphalangeal joints, Di (i = 2–5): finger distal interphalangeal joints. EL: elbow joint (not represented here), WR: wrist joint, RS: radial styloid, US: ulnar styloid, PALM: palm center.

https://doi.org/10.1371/journal.pone.0228869.g001

The following kinematics are considered for the thumb and fingers (thumb: i = 1, index: i = 2, middle: i = 3, ring: i = 4 and pinky: i = 5) with EFi (i = 1–5) that represents the finger and thumb end-effector, Mi (i = 1–5) the finger and thumb metacarpophalangeal joint, Pi (i = 1–5): the finger and thumb proximal interphalangeal joints and Di (i = 2–5) the finger distal interphalangeal joints. Then, the relative orientation of each finger with the hand is represented with a flexion angle αi (Angle between and Mi - Pi projected in the XY plan) and an abduction angle φi (Angle between and Mi - Pi projected in the YZ plane). Moreover, the relative orientation of the thumb with the hand is represented with a flexion angle α1 (Angle between and Mi - Di projected in the YZ plan) and an abduction angle φ1 (Angle between and Mi - Di projected in the XY plane). Furthermore, the angle βi between Pi - Di and Pi - Mi and the angle ϴi between EFi - Di and Pi - Di for the ith finger (i = 2–5) and the angle ϴ1 between EF1 - D1 and M1 - D1 for the thumb are also computed. Finally, dj (j = 1–9) represent respectively the Euclidian distance between each successive EF (EF1 with EF2, EF2 with EF3 …) and between EFi (i = 1–5) and the Palm.

Dataset

Each sign is manually extracted and interpolated to 100 frames with cubic spline interpolation. The dataset is composed of a total number of d = 16890 labeled signs. f = 60 features are computed to represent the kinematics of the right and left sides (Fig 1). For each sign extracted, the features were regrouped in a time-series matrix Fl (m = 100 and n = 60) with each group given by: (1) the finger and thumb end-effector (EF) relative distance, (2) their distance with the palm, (3) the relative orientation of the hand with the forearm, (4) the relative orientation of the fingers and thumb with the hand and (5) the thumb and finger angles. (1) All feature matrices were computed for all signs (l) and subjects (k). Finally, each feature fi,j was standardized with the mean μ and standard σ deviation of their respective group presented above as follow: (2) Then, each subject in the dataset is represented as with N the numbers of sign and yl the corresponding outputs labels of Fl represented as a binary vector.

Data augmentation

To prevent the overfitting of the model and help generalize the classification, data augmentation was considered. A given signal fi is sliced into 10 parts of equal length. Then, data warping was considered on the magnitude and the temporal location of the signal with the following properties. First, each part was distorted to a random value p with a cubic spline interpolation and all parts were reconstructed and interpolated to 100 frames. Then, the amplitude of the signal was slightly altered. A sinus wave with a random period P , phase φ and amplitude A was generated. Finally, this sinus wave was multiplied by A and by the amplitude range of the current signal fi and added to it, thus allowing a smooth variation. This procedure is repeated 40 times for each Fl in the training set (described in the next section).

Gesture classification

Training, validation, and test

This study considers a user-independent approach, i.e. data from the same participant appears either in the train, validation or test sets. To do so, nested k-fold cross-validation with a user-independent approach is considered. It consists of an inner and an outer loop. In each loop, the model is always trained from a random initial point to keep each inner and outer loop independent from each other. Moreover, the first outer loop is considered for hyper-parameters tuning (e.g. number of layers, cells unit size, learning rate, strides, patch size…).

The inner loop consists of splitting the participants into two sets: 20 subjects are assigned to the training/validation set and 5 to the test set. The test set was kept aside to assess the final performance of the models. The training/validation set is used in a non-exhaustive user-independent k-Fold cross-validation that consisted of a rotating K = 4 folds with 15 and 5 subjects to train and validate respectively. To reduce overfitting of the neural network, early-stopping is applied at each k-fold when the accuracy of the validation set starts to decrease (overfitting of the model on the training set).

The outer loop follows the same procedure as the inner loop except that the test set is changed (as well as the train/validation one) with user that still appeared once either in the train/validation or test set. This loop is performed until all participants are tested.

Training properties

Models were trained with mini-batches composed of 500 samples and a learning rate of η = 0.001 with exponential decay of 0.9 every 5 epochs. As a form of regularization, a dropout wrapper is added on each layer to randomly select units that are ignored at each epoch with a probability value of 0.8. The Adam gradient descent optimization algorithm [25] was used to minimize the cost function E that corresponds to a softmax cross-entropy between the estimated vector (logits) (y’) and the true labels vector (y): (3) For each k-fold cross-validation, the training phase was stopped when the accuracy of the validation set starts to decrease and the corresponding model was saved.

After the k-fold cross-validation in each outer loop, 4 trained models are created. The test set is fed to each model (k = 4) and the final predicted class is considered with a majority vote decision.

Models

In this study, different conventional machine learning and deep learning models are considered. Regarding the conventional machine learning models, k-nearest neighbors (k-NN) [26], random forest (RF) [27] and support vector machine (SVM) [28] are used as a baseline comparison. The retained hyperparameters are as follows: k-NN is used with k = 10, RF is used with 1000 trees classifier with a depth of 5 and SVM is used with a linear kernel. The deep learning models considered were Multi-layer Perceptron (MLP), ConvNet, RNN with LSTM cell from Zaremba et al. [29] and DeepConvLSTM (Fig 2). Detailed regarding their hyperparameters is provided in Table 2.

thumbnail
Fig 2. DeepConvLSTM architecture.

ConvNet-LSTM architecture. N1, N2, N3, and N4 represent the numbers of channels (also called feature map) created after each convolutional layer. Their values are respectively 16, 16, 32 and 64. ReLU (rectified linear unit) represent the activation function.

https://doi.org/10.1371/journal.pone.0228869.g002

thumbnail
Table 2. Detail of the hyperparameters retained for Multi-layer Perceptron (MLP), convolutional neural network (ConvNet), recurrent neural network (RNN) and DeepConvLSTM.

https://doi.org/10.1371/journal.pone.0228869.t002

Tensorflow 1.4 [30] with Python 3.5.2 was used and the training was performed on a GTX 1060 6GB that integrates 1280 CUDA cores [31].

Specifically to DeepConvLSTM, the output of the last convolutional layers is fed into 2 LSTM layers composed of the LSTM cell from Zaremba et al. [29]. Moreover, as the time-series is fed to the LSTM layers, the LSTM hidden state contains more and more information about the current sequence. Therefore, the last time step of the LSTM layer is connected to the softmax layer to retrieve the class probabilities. Finally, the last output of the last recurrent layers is connected to a layer with a softmax activation function to normalize the output to a probability distribution.

Results

The results from all different models considered with and without data augmentation are presented in Table 3. To compare the results of the different approaches, a one-way Analysis of Variance (ANOVA) with repeated measures was performed. Then, the Dunnett post hoc test was used to compare the DeepConvLSTM with data augmentation with all other models. The significance level was set at α = 0.05 and the Statistica software (Statsoft, Tulsa, OK, USA) was used. ANOVA on models showed a significant main effect (F(10, 264) = 42.454, p < 0.05) and Dunnett’s test revealed higher value of accuracy for DeepConvLSTM with data augmentation compared to all other models (p < 0.05 for each comparison) except with ConvNet with data augmentation (p > 0.05).

thumbnail
Table 3. Accuracy (Mean (SD)) on the validation and test set for the different models considered (+DA: Include data augmentation).

The results include all inner and outer loops of the nested k-fold cross-validation (* p < 0.05).

https://doi.org/10.1371/journal.pone.0228869.t003

The normalized confusion matrix and a Sankey diagram created using SankeyMATIC (http://sankeymatic.com/) are presented in the supplementary materials S1 and S2 Figs in the supplementary materials respectively.

Moreover, the holdout method with the same DeepConvLSTM hyperparameters without data augmentation was also considered. It simply consists in splitting the complete dataset into two parts randomly selected (80% training and 20% testing) in a stratified way (the same percentage of labels in each set). The holdout method was repeated 10 times and results showed a mean accuracy of 0.997 (0.001) with the test set.

Discussion

The aim of this study was to classify 60 signs of ASL using deep neural networks with a forearm, hand, and finger kinematics models from joint position data provided by the LeapMotion. The raw data from the LeapMotion for the 25 subjects are freely provided with this study (https://doi.org/10.17632/8yyp7gbg6z.1). Moreover, a robust user-independent k-fold cross-validation was used to create the model. This step was composed of four cycles (k = 4) with 15 subjects for training and 5 subjects for validation without overlap. It was followed by a user-independent test phase with 5 subjects used to apply the model in a “real-world case” to present the classification results. Moreover, the DeepConvLSTM was compared with a recurrent neural network composed of the LSTM cell from Zaremba et al. [29] to demonstrate the improvement of a combined approach.

Results showed that the DeepConvLSTM with data augmentation showed the highest results with 91.1 (3.8) % of accuracy on the test sets. The effect of data augmentation on this model showed a significant improvement of about 3.8% (Table 2) demonstrating the importance of using data augmentation technique to help the classifier to generalize human behavior [32]. ConvNet with data augmentation showed non-significant different results compare to DeepConvLSTM with an accuracy of 89.3 (4.0) % and RNN presented a lower accuracy of 87.2 (5.1) % (Table 2). Contrasting with the study of Ordóñez and Roggen [17], the DeepConvLSTM did not outperform ConvNet but outperformed RNN demonstrating the importance of using convolutional layers for HAR. MLP, k-NN, SVM, and RF showed poorer results compared to models composed of convolution or recurrent layers. The current approaches considered the use of kernel size in the convolutional layers along the time axis (e.g. 20 x 1 on the first convolutional layer). Using a larger kernel or changing the input tensor size to 100 x 1 x 60 to get the first convolution to capture all the vector interaction was also tested but was not as successful. It may be difficult for the convolutional layer to capture 60 dimensions along the feature axis at once. Convolutional layers are used to remove outliers’ values, extract features and filter the input data independently along the time axis, making the classification more efficient in the classifier part of the models. RNN models with LSTM cells reach high accuracy (87.2 (5.1) % with data augmentation) which are improved by the convolutional layer in the DeepConvLSTM models. Nevertheless, using a 1D kernel on each feature independently along the time axis in this study provided the best results. As also pointed out by Ordóñez and Roggen [17], using max-pooling at each convolutional layers provided poorer results. Moreover, the test time was about 15.3 (0.6)s and 8.8 (0.3)s. for 5 tested subjects in all outer loop for the DeepConvLSTM and ConvNet respectively providing an advantage of using ConvNet for portable systems.

Regarding the holdout method (used for demonstration purpose only), the accuracy on the validation set was at 99.7 (0.1) %. This result showed how easy it is to gain very high accuracy. This method should be avoided since the models have already learned all the subject’s behavior and the results would remain specific to these subjects and would not be representative of new users. Moreover, other DeepConvLSTM hyperparameters that provided lower accuracy with the user-independent approach also provided accuracy around 99% with the holdout method demonstrating the risk to create models with non-adapted hyperparameters. It remains difficult to compare our results with those of previously reported studies since they used different validation and/or test methods, but we believe that considering a user-independent approach with nested k-fold cross-validation is a relevant way to prove the reliability of a model.

The Sankey diagram (S2 Fig) shows for each true label y the wrong logits y’ that has been identified in an easy-to-read manner. For a better reading, the wrong logits mislabelled 1 and 2 times for each label and subject independently were removed from the Sankey Diagram. Some movements of the sign language were confused because of similarities between them or difficulty from the LeapMotion to correctly detect all joint positions. For instance, Two was confused with Three, Seven with height and Four with Five. Indeed, it was sometimes difficult for the LeapMotion sensor to differentiate the numbers of extended fingers. Movement such as dog (forearm in supination with index finger extended and snap the thumb with the middle finger) is difficult to record by the LeapMotion since the thumb and the middle finger are hidden by the palm and can be confused with one (index finger extended) and brush (index finger extended moving in front in the mouth). Thanks consist of only moving the forearms with the hand flat and finger close to each other while warm consists of the same movement with finger performing and abduction during the movement. The model may receive too limited information to perfectly differentiate both signs. The same problem is observed with car, coat, and cold since they are movements that require moving the forearms with the fists closed which may also provide too limited information given the model here used to be differentiated properly. Finally, Dad was confused with Mom since they required the same hand movement to be performed except that the position in front of the head is different from the thumb position at the chin and at the forehead respectively. Despite this, Dad was confused with Mom for 23% in the test set showing the capability of the DeepConvLSTM to differentiate the slight difference in the global dynamics of the movement. Improvement in the hand tracking with depth-sensor may leverage these problems and would help to reach higher accuracy in the future.

The main limitation of this study is that the recruited participants were new to sign language. Beginners may have greater variability when performing movements and future work should be considered with experts to assess if there would be an improvement in the classification. In addition, a static neural network was used and future work will have to be considered with a dynamic neural network that allows adapting the length of the input sequence coupled with an automatic signal detection method [33]. Nevertheless, a static neural network may still be considered with an automatic segmentation method and then interpolate the signal. Finally, future studies should consider another model that will be fed by the output of the DeepConvLSTM with the purpose of predicting the most probable word based on the previous words/sentences to help the classification from a given context. Finally, sign language recognition may be improved with data provided by a camera or IMU sensors to gain information about the hand position relative to the body.

Conclusion

This study compared various conventional machine learning and deep learning models to classify American sign language. Moreover, a robust user-independent k-fold cross-validation and test phase were provided. This contrast previous work, where the validation and/or the test phase were not user-independent, or lack of information was provided. There are several possibilities for future work to improve these results, such as the use of experts in sign language, dynamic neural network, automatic segmentation technique and additional data from a camera or IMU sensors.

Supporting information

References

  1. 1. Savur C, Sahin F. American Sign Language Recognition system by using surface EMG signal; 2016 9–12 Oct. 2016. pp. 002872–002877.
  2. 2. Sarawate N, Leu MC, ÖZ C (2015) A real-time American Sign Language word recognition system based on neural networks and a probabilistic model. Turkish Journal of Electrical Engineering & Computer Sciences 23: 2017–2123.
  3. 3. Zafrulla Z, Brashear H, Starner T, Hamilton H, Presti P. American sign language recognition with the kinect; 2011. ACM. pp. 279–286.
  4. 4. Kumar P, Gauba H, Roy PP, Dogra DP (2017) Coupled HMM-based multi-sensor data fusion for sign language recognition. Pattern Recognition Letters 86: 1–8.
  5. 5. Dong C, Leu MC, Yin Z. American sign language alphabet recognition using microsoft kinect; 2015. pp. 44–52.
  6. 6. Marin G, Dominio F, Zanuttigh P. Hand gesture recognition with leap motion and kinect devices; 2014 27–30 Oct. 2014. pp. 1565–1569.
  7. 7. Weerasekera C, Jaward MH, Kamrani N. Robust asl fingerspelling recognition using local binary patterns and geometric features; 2013. IEEE. pp. 1–8.
  8. 8. Ravi S, Suman M, Kishore PVV, Kumar EK, Kumar MTK, et al. (2019) Multi modal spatio temporal co-trained CNNs with single modal testing on RGB–D based sign language gesture recognition. Journal of Computer Languages 52: 88–102.
  9. 9. Shin H, Kim WJ, Jang K-a. Korean sign language recognition based on image and convolution neural network; 2019. ACM. pp. 52–55.
  10. 10. Yang S, Zhu Q. Continuous Chinese sign language recognition with CNN-LSTM; 2017. SPIE. pp. 7.
  11. 11. Weichert F, Bachmann D, Rudak B, Fisseler D (2013) Analysis of the accuracy and robustness of the leap motion controller. Sensors 13: 6380–6393. pmid:23673678
  12. 12. Naidu C, Ghotkar A (2016) Hand Gesture Recognition Using Leap Motion Controller. International Journal of Science and Research (IJSR) ISSN (Online): 2319–7064.
  13. 13. Chuan C-H, Regina E, Guardino C. American sign language recognition using leap motion sensor; 2014. IEEE. pp. 541–544.
  14. 14. Fok K-Y, Ganganath N, Cheng C-T, Chi KT. A real-time asl recognition system using leap motion sensors; 2015. IEEE. pp. 411–414.
  15. 15. Lee H, Grosse R, Ranganath R, Ng AY. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations; 2009. ACM. pp. 609–616.
  16. 16. Koller O, Zargaran O, Ney H, Bowden R. Deep Sign: Hybrid CNN-HMM for Continuous Sign Language Recognition; 2016.
  17. 17. Ordóñez FJ, Roggen D (2016) Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16: 115.
  18. 18. Sainath TN, Vinyals O, Senior A, Sak H. Convolutional, long short-term memory, fully connected deep neural networks; 2015. IEEE. pp. 4580–4584.
  19. 19. Zimmermann T, Taetz B (2018) IMU-to-Segment Assignment and Orientation Alignment for the Lower Body Using Deep Learning. 18.
  20. 20. Gers FA, Schmidhuber J, Cummins F (2000) Learning to forget: continual prediction with LSTM. Neural Comput 12: 2451–2471. pmid:11032042
  21. 21. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9: 1735–1780. pmid:9377276
  22. 22. Lipton ZC, Berkowitz J, Elkan C (2015) A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:150600019.
  23. 23. Carpinella I, Jonsdottir J, Ferrarin M (2011) Multi-finger coordination in healthy subjects and stroke patients: a mathematical modelling approach. Journal of neuroengineering and rehabilitation 8: 19. pmid:21507238
  24. 24. Wu G, van der Helm FCT, Veeger HEJ, Makhsous M, Van Roy P, et al. (2005) ISB recommendation on definitions of joint coordinate systems of various joints for the reporting of human joint motion—Part II: shoulder, elbow, wrist and hand. J Biomech 38: 981–992. pmid:15844264
  25. 25. Kingma D, Ba J (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980.
  26. 26. Altman NS (1992) An introduction to kernel and nearest-neighbor nonparametric regression. The American Statistician 46: 175–185.
  27. 27. Ho TK. Random decision forests; 1995. IEEE. pp. 278–282.
  28. 28. Cortes C, Vapnik V (1995) Support-vector networks. Machine learning 20: 273–297.
  29. 29. Zaremba W, Sutskever I, Vinyals O (2014) Recurrent neural network regularization. arXiv preprint arXiv:14092329.
  30. 30. Mart, #237, Abadi n, Barham P, Chen J, et al. (2016) TensorFlow: a system for large-scale machine learning. Proceedings of the 12th USENIX conference on Operating Systems Design and Implementation. Savannah, GA, USA: USENIX Association. pp. 265–283.
  31. 31. Nickolls J, Buck I, Garland M, Skadron K (2008) Scalable Parallel Programming with CUDA. Queue 6: 40–53.
  32. 32. Um TT, Pfister FMJ, Pichler D, Endo S, Lang M, et al. (2017) Data augmentation of wearable sensor data for parkinson&#8217;s disease monitoring using convolutional neural networks. Proceedings of the 19th ACM International Conference on Multimodal Interaction. Glasgow, UK: ACM. pp. 216–220.
  33. 33. Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: Unified, real-time object detection; 2016. pp. 779–788.