Next Article in Journal
Parameterized Design and Dynamic Analysis of a Reusable Launch Vehicle Landing System with Semi-Active Control
Previous Article in Journal
Prospective Teachers’ Development of Meta-Cognitive Functions in Solving Mathematical-Based Programming Problems with Scratch
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Hand-Oriented Activity Recognition Based on Smartwatch Sensor Data Using LSTMs

by
Sakorn Mekruksavanich
1,
Anuchit Jitpattanakul
2,*,
Phichai Youplao
3,4,* and
Preecha Yupapin
3,4
1
Department of Computer Engineering, School of Information and Communication Technology, University of Phayao, Phayao 56000, Thailand
2
Intelligent and Nonlinear Dynamic Innovations Research Center, Department of Mathematics, Faculty of Applied Science, King Mongkut’s University of Technology North Bangkok, Bangkok 10800, Thailand
3
Computational Optics Research Group, Advanced Institute of Materials Science, Ton Duc Thang University, District 7, Ho Chi Minh City 729000, Vietnam
4
Faculty of Applied Sciences, Ton Duc Thang University, District 7, Ho Chi Minh City 729000, Vietnam
*
Authors to whom correspondence should be addressed.
Submission received: 29 August 2020 / Revised: 14 September 2020 / Accepted: 18 September 2020 / Published: 22 September 2020
(This article belongs to the Section Computer)

Abstract

:
The creation of the Internet of Things (IoT), along with the latest developments in wearable technology, has provided new opportunities in human activity recognition (HAR). The modern smartwatch offers the potential for data from sensors to be relayed to novel IoT platforms, which allow the constant tracking and monitoring of human movement and behavior. Recently, traditional activity recognition techniques have done research in advance by choosing machine learning methods such as artificial neural network, decision tree, support vector machine, and naive Bayes. Nonetheless, these conventional machine learning techniques depend inevitably on heuristically handcrafted feature extraction, in which human domain knowledge is normally limited. This work proposes a hybrid deep learning model called CNN-LSTM that employed Long Short-Term Memory (LSTM) networks for activity recognition with the Convolution Neural Network (CNN). The study makes use of HAR involving smartwatches to categorize hand movements. Using the study based on the Wireless Sensor Data Mining (WISDM) public benchmark dataset, the recognition abilities of the deep learning model can be accessed. The accuracy, precision, recall, and F-measure statistics are employed using the evaluation metrics to assess the recognition abilities of LSTM models proposed. The findings indicate that this hybrid deep learning model offers better performance than its rivals, where the achievement of 96.2% accuracy, while the f-measure is 96.3%, is obtained. The results show that the proposed CNN-LSTM can support an improvement of the performance of activity recognition.

Graphical Abstract

1. Introduction

Miniature sensors have driven rapid growth in the wearable technology sector, since they permit computerized monitoring at any time [1]. A wide range of affordable wearable products are now available, including smartwatches, smartphones, and smart glasses. Through their internal gyroscopes and accelerometers, these devices are able to record significant quantities of data which can then undergo analysis to determine the various activities the wearer has performed [2]. It is then possible to classify those activities and use the information for various purposes, such as monitoring the elderly [3], tracking exercise [4], or protecting people from office workers’ syndrome [5].
To date, mobile computing and human activity recognition have become popular and complex [6,7]. The data obtained from the various wearable sensors can be treated as time-series data, where a majority of studies have focused on the recognition of activity through the use of smartphones [8,9,10,11]. Almost everyone now carries a smartphone, which increases the potential of the applications they use. However, one drawback is the fact that phones are usually carried in bags or pockets, are therefore not consistently positioned on the body. Furthermore, a phone held in the trouser pocket is not ideally positioned to record hand movements, which cannot even be relied upon to maintain the same orientation as it is carried around. For female users, the smartphone is especially inconvenient as many do not even carry it in their pockets, preferring instead to use a handbag. Smartwatches, however, can solve many of these issues, and since they are worn in a fixed position on the wrist, they are the best choice to capture hand movements [12].
There are a number of benefits offered by smartwatches, in which other forms of inertial sensor technology cannot match. Among these include the fact that smartwatches are able to merge the features of a smartphone with the ability to monitor data constantly. Through small screens, it is possible for the user to interact with the smartwatch wherever they are located. The watch can be worn at all times, will be waterproof, and typically has a much longer useful battery lifespan than a smartphone [13]. Moreover, smartphones cannot always be carried during periods of strenuous exercise due to their size and shape. In contrast, smartwatches can be worn while exercising and can, therefore, generate data from the compass, GPS, accelerometer, and gyroscope, as well as monitoring heart rates [14,15].
Today’s greater availability of wearable technology has brought about increased interest in HAR in order to bring about benefits for people’s health and well-being [16]. During the past five years there has been a notable increase in the number of research papers published in this particular field [17], with a majority of these studies focusing on the applications of HAR in conventional machine learning (ML) models. Such models typically employ ML algorithms such as support vector machines, naive Bayes, decision trees, hidden Markov, and k-nearest neighbor models. The drawback to such approaches, however, is the reliance upon hand-crafted shallow feature extraction, which is a practice that relies heavily upon the knowledge of the humans performing the work [18]. Furthermore, these ML approaches apply independent statistical formulas to segment each time step within the time-series data, which leads to the time and space relationships within the data being lost when the models are trained. However, in more recent years, published studies have begun to show the accomplishments of deep learning (DL) models in addressing complicated HAR issues [19,20]. These techniques differ from the traditional ML approaches since it becomes possible to address high-level features when examining raw data from sensors.
In this study, a smartwatch-based approach to HAR is examined in the context of hand movement recognition using data from sensors in a smartwatch, along with the Wireless Sensor Data Mining (WISDM) public smartwatch dataset, which is composed of IMU sensor data that records the activity of the hands as well as data associated with eating of 51 participants were provided with the data. This study suggested the use of a hybrid deep learning model that employed LSTM networks for recognition performance improvement.
The remaining paper is structured as follows. Section 2 presents the preliminary background and related works in the research work of human activity recognition. In Section 3, an overview of the proposed smartwatch based HAR framework is presented. The details of each process in the framework are described. Section 4 conducts the experiments from various scenarios, as well as showing improvable results by the proposed framework comparing with state-of-the-art evaluation metrics. Finally, Section 5 presents the research conclusion and challenging future points.

2. Preliminary Background

2.1. Human Activity Recognition

The definition of human activity covers all physical movements involved in human life, whether necessary or otherwise. Such activity requires energy to be spent in moving the skeletal muscles, with examples including walking, eating, sitting, standing, and so forth. When conducting research involving time-series classification, human activity recognition (HAR) faces the challenges of classifying the sequences of data obtained from wearable sensors in time-series format, to determine clearly defined human movements [21,22,23,24].
Suppose a human performs some activities according to a pre-defined activity set A = { a 1 , a 2 , a 3 , …, a m }, where m means the number categories of physical activity. Given a time-series sequence s = { d 1 , d 2 , …, d t , …, d n } of sensor data reading is captured from the human activity information, where d t means the data sensor reading at time t and n denoted of sequence and n ≥ m. The HAR task aims to determine a function F to forecast the physical activity sequence which based on data sensor reading s. The function F can be defined:
F ( s ) = { a 1 , a 2 , a 3 , , a n } , a n A ,
Meanwhile the real activity sequence is indicated as:
F ( s ) = { a 1 , a 2 , a 3 , , a n } , a n A ,
The general framework encompassing HAR using wearable sensor data is shown in Figure 1. This framework comprises four key processes which can be employed to categorize the various forms of physical activity based on raw data from the activity sensors. In the first step, the raw data are gathered from the IMU sensors carried in wearable technology. Then in step two, the data are cleaned and prepared through the preprocessing phase [25]. This involves filtering and addressing the issue of missing data from sensor errors by including estimations based on statistical formulas. The data must then be segmented and balanced. However, because the sensor data appears in time-series form, the characteristic features is necessary to be extracted from the high-dimensional data. In the final stage, these features will then be used to train the model through the use of various ML algorithms including naive Bayes, support vector machines, or decision tree. Once trained, the ML model will then be capable of recognizing and categorizing human activity.

2.2. Deep Learning with HAR

Important recent studies of HAR [17] have revealed certain problems associated with conventional machine learning techniques which ultimately influence the ability to recognize human activity. This limitation concerns the choice of hand-crafted features since the selection is dependent upon the skills and knowledge of the person taking the decisions [26]. Within the time span of just a few years, however, deep learning has been applied as a suitable alternative approach which can address these limitations effectively.
Figure 2 shows the deep learning process can eliminate the need for feature extraction in the general HAR framework. Recently, the studies have put forward several models for deep learning, which can be addressed the issues related to the time-series classification of HAR problems. Studies have examined some learning models to assess their recognition abilities in terms of miscellaneous benchmark activity datasets. Two of the main models are Convolution Neural Networks (CNNs) and Long Short-Term Memory Networks (LSTMs); they are effective for HAR problems using smartphones since they offer appropriate evaluation metrics. These two models are thus investigated in this study to determine their effectiveness in recognizing hand movements using data from smartwatches.

3. Proposed Framework

In this section, the proposed framework for hand-oriented activity recognition based on smartwatch sensor data is presented. The overall process of the framework is depicted as Figure 3.
The framework consists of four main processes (Data collection, Data preprocessing and segmentation, Deep model training, and Model evaluation).

3.1. Data Collection

The smartwatch data used in the research work is a public benchmark dataset called WISDM from the UCI Repository [27]. This dataset provides tri-axial accelerometer data and tri-axial gyroscope data collected at a rate of 20 Hz from Android smartphones and an Android smartwatch. Data is gathered at a rate of 20 Hz in every 50 ms. The Android smartphone and smartwatch are Samsung Galaxy S5 with Android 6.0 operating system and LG G Watch running with Android Wear 1.5, respectively. These raw sensor data are recorded from 51 subjects with 18 and 25 years who performed 18 pre-defined physical activities. All subjects wear the smartwatch on their dominant hand while they are performing the activities [28,29]. These physical activities can be categorized into three main categories, i.e., non-hand-oriented activities, hand-oriented eating activities, and hand-oriented general activities. Some samples from accelerometer data and gyroscope data are illustrated in Figure 4 and Figure 5, respectively.

3.2. Data Preprocessing and Segmentation

The activities from the WISDM dataset have been found to be recorded separately from the exploratory data analysis. This means that each person performs one activity for approximately 3 min in a raw data, followed by the next activity, etc. When looking at the timestamps, it can be seen that the transitions from one activity to the next one are not continuous, but the recording has happened in isolation. The smartphone and smartwatch data is not also synchronized, i.e., they have not been collected in parallel.
This research analyzes the sensor data by exploratory data analysis and finds that there are activity data of 7 subjects did not contain all pre-defined activities. So, the seven subjects are cut-off, i.e., subject number 1616, 1618, 1637, 1638, 1639, 1640, and 1642. To handle time-series data on the HAR problem with 75% of the overlapping proportion as shown in Figure 6, the smartwatch data remains with 44 subjects from 55 subjects, where the data is also segmented by a sliding window of 10 s. The number of raw sensor data is shown in Table 1.
Given the sampling rate of 20 Hz and the windows size of 10 s, there are 200 raw sensor data for each sample. Then these sensor samples are split to 70% as training set and remaining 30% as test set for evaluating the proposed model. The numbers of smartwatch sensor samples used in this work is presented in Table 2.

3.3. Deep Model Training

LSTM is an extended deep learning model of recurrent neural networks (RNN) [30]. LSTM is proposed to tackle vanishing and exploding gradient problems [31]. The computational process of the LSTM network can be detailed as the following process: (1) any input data X = { x 0 , x 1 , x 2 , …, x t , x t + 1 , …} is turned into the data together with the data of the hidden layer at the last stage to the hidden layer H = { h 0 , h 1 , h 2 , …, h t , h t + 1 , …} by the matrix transformation technique, and (2) the hidden layer output passes under an activation function to the last value of the output layer Y = { y 0 , y 1 , y 2 , …, y t , y t + 1 , …} as illustrated in Figure 7.
The LSTM network consists of the specific characteristic structure for memorizing information for a stretched time. The input and forget gate manipulate process to replace the new result information by comparing the internal memory with the new data feeding. This process enhances gradients to circulate effectively in time.
As shown in Figure 8, input gate (i), forget gate (f), output gate (o) and memory cell (C) of LSTM are designed to control what information should be stored, updated, and deleted. Gating is the technique to pass the needed information selectively [32]. This technique composes of the Sigmoid and the Hadamard product function. Derived output value within range of [0, 1] is allowed the multiplication to proceed the information. This consideration is the effective practice to initialize these gates to a value of 1 or close to 1. So, this situation is not to debilitate training initially. Therefore, each parameter at the moment t can be defined in the LSTM node.
From Figure 8, each gate that works inside an LSTM cell is mathematically defined as the following expressions [33]:
f t = σ ( U f x t + W f h t 1 + b f )
i t = σ ( U f x t + W f h t 1 + b f )
g t = tanh ( U g x t + W g h t 1 + b c )
o t = σ ( U o x t + W o h t 1 + b o )
c t = f t c t 1 i t g t
h t = o t tanh ( c t )
where:
  • Forget gate ( f t ) manipulates when there is desired unnecessary data to delete.
  • Input gate ( i t ) manipulates the input activation of new data into the memory cell.
  • Input modulation gate ( g t ) manipulates the preeminent input to the memory cell.
  • Output gate ( o t ) manipulates the flow of output.
  • Internal state ( c t ) manipulates the constitutional recurrence of cell.
  • Hidden state ( h t ) manipulates the data from the preceding data case inward the context window.
Equations (3)–(8) described the detailed process of LSTM network as following: (1) When there is a need to delete unnecessary data that involves the forget gate. (2) New data which needs to store in memory with an input gate is determined. Then the state value of an old node C t 1 is updated to the new node state C t . (3) Finally, which data should be output to the layer over with an output gate is determined.
The hybrid LSTM network called CNN-LSTM is proposed to enhance the recognition performance in this research. The CNN-LSTM is consisted of two convolutional layers with one LSTM layers. The structure of CNN-LSTM can be illustrated in Figure 9. The proposed CNN-LSTM model introduces by 1D-CNN two layers with 3 × 1 kernels for extracting deep spatial appearances from smartwatch sensor data. The number of filters is set to 128 and 64 for the first and second 1D-CNN layer, respectively. The rectified linear function is used as the activation function in the proposed model. After the second 1D-CNN layer is processed, the dropout technique is applied. A LSTM layer with 128 units is then applied before a dropout layer for extracting temporal features. Finally, a completely connected dense layer is employed to the LSTM output at the final time step with a SoftMax function. Figure 10 shows accuracy and loss examples of the training process of CNN, LSTM, and the proposed CNN-LSTM. The gap between training and validation accuracy is a clear indication of the overfitting problem. Many gaps in the network affect to increase overfitting problems.

3.4. Model Evaluation

Generally four evaluation metrics for multi-class classification are employed to verify the performance of the proposed CNN-LSTM models. Accuracy, precision, recall, and F-measure are considered as performance metrics evaluation in this research work. When correctly recognized, such activity could be classified as True Positive (TP) or True Negative (TN). On the other hand, activity can be classified as False Positive (FP) or False Negative (FN) when incorrectly classified by metrics. True Positive or True Negative derive other performance metrics. Given TP = c = 1 N TP c denotes the total true positive rate for a classifier on all classes, TN = c = 1 N TN c denotes the total true negative rate, FP = c = 1 N FP c denotes the total false positive rate, and FN = c = 1 N FN c denotes the total false negative rate. These metrics are detailed following.

3.4.1. Accuracy

Accuracy shows that ratio of correctly classified issues which is the sum of correct classification divide by the total number of classifications.
Accuracy = TP + TN TP + FP + TN + FN

3.4.2. Precision

Similar to accuracy, precision measures the accuracy value based on the negative instance fraction, which is classified as negative. The total precision (Precision) is the average of the precision for each class mean by Precisionc:
Precisionn c = TP c TP c + FP c
Precision = 1 N c = 1 N TP c TP c + FP c

3.4.3. Recall

Recall shows the effectiveness of correctly forecasted issues as positive instances. The total recall (Recall) is the average of the recall for each class mean by Recallc:
Recall c = TP c TP c + FN c
Recall = 1 N c = 1 N TP c TP c + FN c

3.4.4. F-Measure

F-measure helps to search the balancing of precision and recall. There is an uneven class distribution (larger amount of actual negatives). The F-measure for each class is denoted by F-measurec.
F-measure c = 2 × Recall c × Precision c Recalll c + Precision c
F-measure = 2 × Recall × Precision Recall + Precision

4. Experimental Results

4.1. Experiments

Experimental implementation is conducted by using Python 3.6.9, TensorFlow 2.2.0, Keras 2.3.1, Scikit-Learn, Numpy 1.18.5, and Pandas 1.0.5 libraries in this research. The experiments have been executed on the Google Colab platform with Tesla K80 and also optimized hyperparameters of each deep learning model by SigOpt. Three different experiments are set up with three DL models—CNN, LSTM, and proposed CNN-LSTM—as following scenarios:
  • Scenario 1—Only raw accelerometer data,
  • Scenario 2—Only raw gyroscope data,
  • Scenario 3—Raw accelerometer data and raw gyroscope data.
Table 3 shows layer types of baseline deep learning models (CNN anf LSTM) with the proposed CNN-LSTM model.

4.2. Experimental Results

For performance evaluation, smartwatch sensor data is used from WISDM dataset by splitting 70% for training models and the remaining 30% for testing the trained models. The experimental results of all three scenarios are shown in Table 4.
From the results in Table 4, the proposed CNN-LSTM model outperforms the other DL models in every activity. Though when analyzing the performance metrics recognition with accuracy, precision, recall, and F-measure, results of F-measure metric is only presented because they incorporate accuracy and precision. The confusion matrix for the proposed CNN-LSTM networks is shown in Table 5. Because of focusing on the comparison of different activities in this research, F-measure is chosen as the performance metric for this objective. Moreover, we can compare the effectiveness of the CNN-LSTM with the F-measure of each category of activities. The F-measure results of the different DL models trained from smartwatch sensor data are shown in Figure 11.
In order to avoid that results are biased by the particular choice of training and test dataset, the second experiment is conducted by utilizing different combination. The training, validation, and test sets are obtained using stratified ten-fold cross validation. As for the subsample, 80% of them, are used as training data. The remaining 20% are split for validation and test sets. In each fold, we ensure that the subsamples from each trial are not assigned to more than one set. The experimental results of stratified ten-fold cross validation with standard deviation value is shown in Table 6.

4.3. Comparison of Proposed CNN-LSTM Network and Other Conventional Deep Learning Networks

The significance of some deep learning network structures on model operation is explored in this section. As detailed shown in Table 4, three types of model compositions—CNN, LSTM, and the proposed CNN-LSTM—are implemented respectively for experimental results comparison. Activity recognition results are evaluated by the advanced metrics such as accuracy, precision, recall, and F-measure on the test collection. Moreover, the computation speed of training iteration is given. The research experiments are conducted by depended on the WISDM dataset.
The classical convolutional neural network (CNN) structure is commonly followed by a totally-connected layer to incorporate the features extracted from the previous layer. As can be seen that the F-measure of the CNN attends to 93.10% when the trained model is revealed to the test set composing with both accelerometer and gyroscope sensor data. In spite of classical CNN can enhance the recognition performance of the model, a massive number of parameters which more than 4 million model parameters have unavoidably to be carried. Moreover, the training stage time per epoch is still more than 9.21 s. The LSTM structure belongs to the vanilla long short-term memory neural network. This structure gains the capability to perceive temporal information from sequential data. The F-measure result of this model comes to 89.60% when both accelerometer and gyroscope sensor data used for training the model. Different from CNN structure, LSTM structure brings a smaller number of parameters—that are only 52 thousand parameters. The training stage time consumes more than 212.77 s per epoch.
In the proposed CNN-LSTM structure, the sensor data captured from smartwatch are firstly delivered into two layers of CNNs and then transferred to LSTM layers for performing the feature extraction process. Certainly, the proposed model surpasses in F-measure performance from the other deep learning networks with 96.30%. Moreover, the proposed CNN-LSTM brings a small number of parameter—that are 1.6 million model parameters. Furthermore, the model training time is spent only 20.46 s per epoch. So, the proposed model in this work is not only achieved high recognition rate of F-measure metric, but it also greatly simplifies the model structure with a smaller number of parameters.
Considering to model training speed, Figure 10 displays training speed comparison (in term of model loss and model accuracy). It becomes apparent that training loss from the proposed CNN-LSTM is decreasing faster than the baseline CNN and LSTM model. Moreover, the CNN-LSTM is more suitable than the other conventional deep learning models in term of training speed by time training per epoch.
The previous work [27] uses hand-crafted features on conventional machine learning methods (decision tree and k-nearest neighbor) or an ensemble learning (random forest). The results of this work indicated that the random forest (RF) is more outperform than others. So, the proposed CNN-LSTM model is used to compared with existing trained models by the same smartwatch dataset. The comparative results is summarized in Table 7. The proposed CNN-LSTM can improve the accuracy in every scenario in the combination of sensor data. Generally, the proposed CNN-LSTM model outperform others in the conducted experiments. It is because CNN-LSTM has efficiency at dealing with complicated time series problems. Comparison performance results of the eating activity and general activity between the previous work and the proposed CNN-LSTM model can be shown in Figure 12.

5. Conclusions and Future Works

This work studied hand-oriented activity recognition using sensor data captured from a smartwatch that is collected in the WISDM dataset. The framework for smartwatch-based human activity recognition is proposed to enhance the recognition performance with a hybrid Long Short-Term Memory network. Three scenarios were employed for each deep learning model performance using the different smartwatch sensors. The evaluation metrics such as accuracy, precision, recall, and F-measure were applied. The experimental results show that the proposed CNN-LSTM model outperforms other baseline deep learning models with the highest performance metrics in every scenario. Moreover, the proposed model is compared to other previous works. The comparative results show that the recognition performance of every scenario can be enhanced by the proposed model, especially for hand-oriented activities.
The research of conventional deep learning techniques is generally depended on CNNs and LSTMs in HAR environment. With deep learning CNNs, there are two advantages over other models—-local dependency and scale invariance. The CNNs are nonetheless memory networks that be able to only extract spatially deep features from raw sensor data. Moreover, with another conventional deep learning model—LSTMs, these networks can employ the temporal dependencies in raw sensor data appear as the common alternative for human-based movement modelling with captured sensor data. So, with the proposed method CNN-LSTM in this research, an advantage is relied on to aid in extracting both the spatial features and temporal features of the signal data of human activities. For future work, developing other kinds of deep learning models and improving recognition performance through data augmentation techniques are considered for personalized smartwatch based HAR.

Author Contributions

Conceptualization and model analysis, S.M.; resource and data curation, A.J.; methodology and validation, S.M.; data visualization and graphic improvement, P.Y. (Phichai Youplao); discussion and final editing, P.Y. (Preecha Yupapin); writing-review and editing, S.M.; funding acquisition, A.J. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by King Mongkut’s University of Technology North Bangkok, grant number KMUTNB-62-KNOW-25 and was partially funded by School of Information and Communication Technology, University of Phayao.

Acknowledgments

The authors thank the SigOpt team for the provided optimization services.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jobanputra, C.; Bavishi, J.; Doshi, N. Human Activity Recognition: A Survey. Procedia Comput. Sci. 2019, 155, 698–703. [Google Scholar] [CrossRef]
  2. Vashist, S.K.; Luong, J.H. Chapter 12—Wearable Technologies for Personalized Mobile Healthcare Monitoring and Management. In Wearable Technology in Medicine and Health Care; Tong, R.K.Y., Ed.; Academic Press: Cambridge, MA, USA, 2018; pp. 235–259. [Google Scholar] [CrossRef]
  3. Hnoohom, N.; Jitpattanakul, A.; Inluergsri, P.; Wongbudsri, P.; Ployput, W. Multi-sensor-based fall detection and activity daily living classification by using ensemble learning. In Proceedings of the 2018 International ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI-NCON), Chiang Rai, Thailand, 25–28 February 2018; pp. 111–115. [Google Scholar]
  4. Mekruksavanich, S.; Jitpattanakul, A. Exercise Activity Recognition with Surface Electromyography Sensor using Machine Learning Approach. In Proceedings of the 2020 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT NCON), Pattaya, Thailand, 11–14 March 2020; pp. 75–78. [Google Scholar]
  5. Mekruksavanich, S.; Hnoohom, N.; Jitpattanakul, A. Smartwatch-based sitting detection with human activity recognition for office workers syndrome. In Proceedings of the 2018 International ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI-NCON), Chiang Rai, Thailand, 25–28 February 2018; pp. 160–164. [Google Scholar]
  6. Fu, B.; Damer, N.; Kirchbuchner, F.; Kuijper, A. Sensing Technology for Human Activity Recognition: A Comprehensive Survey. IEEE Access 2020, 8, 83791–83820. [Google Scholar] [CrossRef]
  7. Ferrari, A.; Micucci, D.; Mobilio, M.; Napoletano, P. On the Personalization of Classification Models for Human Activity Recognition. IEEE Access 2020, 8, 32066–32079. [Google Scholar] [CrossRef]
  8. Rahim, K.; Elamvazuthi, I.; Izhar, L.; Capi, G. Classification of Human Daily Activities Using Ensemble Methods Based on Smartphone Inertial Sensors. Sensors 2018, 18, 4132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Ahmed, N.; Rafiq, J.I.; Islam, M.R. Enhanced Human Activity Recognition Based on Smartphone Sensor Data Using Hybrid Feature Selection Model. Sensors 2020, 20, 317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Lu, W.; Fan, F.; Chu, J.; Jing, P.; Yuting, S. Wearable Computing for Internet of Things: A Discriminant Approach for Human Activity Recognition. IEEE Internet Things J. 2019, 6, 2749–2759. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Zhang, Z.; Zhang, Y.; Bao, J.; Zhang, Y.; Deng, H. Human Activity Recognition Based on Motion Sensor Using U-Net. IEEE Access 2019, 7, 75213–75226. [Google Scholar] [CrossRef]
  12. Weiss, G.M.; Timko, J.L.; Gallagher, C.M.; Yoneda, K.; Schreiber, A.J. Smartwatch-based activity recognition: A machine learning approach. In Proceedings of the 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), Las Vegas, NV, USA, 24–27 February 2016; pp. 426–429. [Google Scholar]
  13. Porzi, L.; Messelodi, S.; Modena, C.M.; Ricci, E. A Smart Watch-Based Gesture Recognition System for Assisting People with Visual Impairments. In IMMPD ’13: Proceedings of the 3rd ACM International Workshop on Interactive Multimedia on Mobile and Portable Devices, Barcelona Spain, 22 October 2013; Association for Computing Machinery: New York, NY, USA, 2013; pp. 19–24. [Google Scholar] [CrossRef]
  14. Casilari, E.; Álvarez Marco, M.; García-Lagos, F. A Study of the Use of Gyroscope Measurements in Wearable Fall Detection Systems. Symmetry 2020, 12, 649. [Google Scholar] [CrossRef] [Green Version]
  15. Liu, Y.; Ouyang, D.; Liu, Y.; Chen, R. A Novel Approach Based on Time Cluster for Activity Recognition of Daily Living in Smart Homes. Symmetry 2017, 9, 212. [Google Scholar] [CrossRef]
  16. Burmaoglu, S.; Trajkovik, V.; Tutukalo, T.L.; Yalcin, H.; Caulfield, B. Chapter 14—Evolution Map of Wearable Technology Patents for Healthcare Field. In Wearable Technology in Medicine and Health Care; Tong, R.K.Y., Ed.; Academic Press: Cambridge, MA, USA, 2018; pp. 275–290. [Google Scholar] [CrossRef]
  17. Tüfek, N.; Özkaya, O. A Comparative Research on Human Activity Recognition Using Deep Learning. In Proceedings of the 2019 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; pp. 1–4. [Google Scholar]
  18. Baldominos, A.; Cervantes, A.; Sáez, Y.; Isasi, P. A Comparison of Machine Learning and Deep Learning Techniques for Activity Recognition using Mobile Devices. Sensors 2019, 19, 521. [Google Scholar] [CrossRef] [Green Version]
  19. Murad, A.; Pyun, J.Y. Deep Recurrent Neural Networks for Human Activity Recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Lv, T.; Wang, X.; Jin, L.; Xiao, Y.; Song, M. Margin-Based Deep Learning Networks for Human Activity Recognition. Sensors 2020, 20, 1871. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Kim, H.; Kim, I. Human Activity Recognition as Time-Series Analysis. Math. Probl. Eng. 2015, 2015, 1–9. [Google Scholar] [CrossRef] [Green Version]
  22. Li, F.; Shirahama, K.; Nisar, M.A.; Huang, X.; Grzegorzek, M. Deep Transfer Learning for Time Series Data Based on Sensor Modality Classification. Sensors 2020, 20, 4271. [Google Scholar] [CrossRef] [PubMed]
  23. Kim, E.; Helal, S.; Cook, D. Human Activity Recognition and Pattern Discovery. IEEE Pervasive Comput. 2010, 9, 48–53. [Google Scholar] [CrossRef] [Green Version]
  24. Bulling, A.; Blanke, U.; Schiele, B. A Tutorial on Human Activity Recognition Using Body-Worn Inertial Sensors. ACM Comput. Surv. 2014, 46. [Google Scholar] [CrossRef]
  25. Barth, J.; Oberndorfer, C.; Pasluosta, C.; Schülein, S.; Gaßner, H.; Reinfelder, S.; Kugler, P.; Schuldhaus, D.; Winkler, J.; Klucken, J.; et al. Stride Segmentation during Free Walk Movements Using Multi-Dimensional Subsequence Dynamic Time Warping on Inertial Sensor Data. Sensors 2015, 15, 6419–6440. [Google Scholar] [CrossRef]
  26. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  27. Weiss, G.M.; Yoneda, K.; Hayajneh, T. Smartphone and Smartwatch-Based Biometrics Using Activities of Daily Living. IEEE Access 2019, 7, 133190–133202. [Google Scholar] [CrossRef]
  28. Mahloko, L.; Adebesin, F. A Systematic Literature Review of the Factors that Influence the Accuracy of Consumer Wearable Health Device Data. In Responsible Design, Implementation and Use of Information and Communication Technology; Hattingh, M., Matthee, M., Smuts, H., Pappas, I., Dwivedi, Y.K., Mäntymäki, M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 96–107. [Google Scholar]
  29. Cosoli, G.; Spinsante, S.; Scalise, L. Wrist-worn and chest-strap wearable devices: Systematic review on accuracy and metrological characteristics. Measurement 2020, 159, 107789. [Google Scholar] [CrossRef]
  30. Ordóñez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Devanne, M.; Papadakis, P.; Nguyen, S.M. Recognition of Activities of Daily Living via Hierarchical Long-Short Term Memory Networks. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 3318–3324. [Google Scholar]
  32. DiPietro, R.; Hager, G.D. Chapter 21—Deep learning: RNNs and LSTM. In Handbook of Medical Image Computing and Computer Assisted Intervention; Zhou, S.K., Rueckert, D., Fichtinger, G., Eds.; Academic Press: Cambridge, MA, USA, 2020; pp. 503–519. [Google Scholar] [CrossRef]
  33. Asif, A.R.; Waris, M.; Gilani, S.; Jamil, M.; Ashraf, H.; Shafique, M.; Niazi, I. Performance Evaluation of Convolutional Neural Network for Hand Gesture Recognition Using EMG. Sensors 2020, 20, 1642. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. The general framework of human activity recognition (HAR) using machine learning approaches.
Figure 1. The general framework of human activity recognition (HAR) using machine learning approaches.
Symmetry 12 01570 g001
Figure 2. The general framework of HAR using deep learning approaches.
Figure 2. The general framework of HAR using deep learning approaches.
Symmetry 12 01570 g002
Figure 3. The proposed smartwatch-based HAR framework.
Figure 3. The proposed smartwatch-based HAR framework.
Symmetry 12 01570 g003
Figure 4. Activity accelerometer data for 10 s of (a) kicking, (b) stairs, (c) standing, (d) jogging, (e) sitting, (f) walking, (g) sandwich, (h) chips, (i) drinking, (j) soup, (k) pasta, (l) folding, (m) typing, (n) teeth, (o) catch, (p) clapping, (q) dribbling, and (r) writing.
Figure 4. Activity accelerometer data for 10 s of (a) kicking, (b) stairs, (c) standing, (d) jogging, (e) sitting, (f) walking, (g) sandwich, (h) chips, (i) drinking, (j) soup, (k) pasta, (l) folding, (m) typing, (n) teeth, (o) catch, (p) clapping, (q) dribbling, and (r) writing.
Symmetry 12 01570 g004
Figure 5. Activity gyroscope data for 10 s of (a) kicking, (b) stairs, (c) standing, (d) jogging, (e) sitting, (f) walking, (g) sandwich, (h) chips, (i) drinking, (j) soup, (k) pasta, (l) folding, (m) typing, (n) teeth, (o) catch, (p) clapping, (q) dribbling, and (r) writing.
Figure 5. Activity gyroscope data for 10 s of (a) kicking, (b) stairs, (c) standing, (d) jogging, (e) sitting, (f) walking, (g) sandwich, (h) chips, (i) drinking, (j) soup, (k) pasta, (l) folding, (m) typing, (n) teeth, (o) catch, (p) clapping, (q) dribbling, and (r) writing.
Symmetry 12 01570 g005
Figure 6. Data segmentation process by a sliding window of 10 s to handle time-series data on the HAR problem.
Figure 6. Data segmentation process by a sliding window of 10 s to handle time-series data on the HAR problem.
Symmetry 12 01570 g006
Figure 7. The unfold structure of one-layer baseline Long Short-Term Memory (LSTM).
Figure 7. The unfold structure of one-layer baseline Long Short-Term Memory (LSTM).
Symmetry 12 01570 g007
Figure 8. The structure of an LSTM node.
Figure 8. The structure of an LSTM node.
Symmetry 12 01570 g008
Figure 9. The proposed Convolution Neural Network (CNN)-LSTM network structure in this research.
Figure 9. The proposed Convolution Neural Network (CNN)-LSTM network structure in this research.
Symmetry 12 01570 g009
Figure 10. Accuracy and loss examples of the training process of (a) CNN, (b) LSTM, and (c) proposed CNN-LSTM.
Figure 10. Accuracy and loss examples of the training process of (a) CNN, (b) LSTM, and (c) proposed CNN-LSTM.
Symmetry 12 01570 g010
Figure 11. F-measure of the different deep learning (DL) models (trained from smartwatch sensor data on the WIDSM dataset) of (a) Non-Hand-Oriented Activities, (b) Hand-Oriented Activities (General), and (c) Hand-Oriented Activities (Eating).
Figure 11. F-measure of the different deep learning (DL) models (trained from smartwatch sensor data on the WIDSM dataset) of (a) Non-Hand-Oriented Activities, (b) Hand-Oriented Activities (General), and (c) Hand-Oriented Activities (Eating).
Symmetry 12 01570 g011
Figure 12. Performance (accuracy %) of comparison results: (a) Comparison results of the eating activity, and (b) Comparison results of the general activity
Figure 12. Performance (accuracy %) of comparison results: (a) Comparison results of the eating activity, and (b) Comparison results of the general activity
Symmetry 12 01570 g012
Table 1. Numbers of raw sensor data.
Table 1. Numbers of raw sensor data.
51 SubjectsSelected 44 Subjects
raw accelerometer data3,777,0462,969,030
raw gyroscope data3,440,3422,960,050
Table 2. Numbers of sensor samples
Table 2. Numbers of sensor samples
51 SubjectsTraining Set (70%)Test Set (30%)
raw accelerometer data59,38041,56617,814
raw gyroscope data59,20141,44117,760
Table 3. Layer types of baseline deep learning models and the proposed CNN-LSTM mode.
Table 3. Layer types of baseline deep learning models and the proposed CNN-LSTM mode.
DL ModelsLayer Types
CNNConv1D–Dropoput–Maxpooling–Dense–Softmax
LSTMLstm–Dropout–Dense–Softmax
CNN-LSTMConv1D–Conv1d–Dropout–Maxpooling–Lstm–Dropout–Dense–Softmax
Table 4. Experimental results of all three scenarios with two baseline deep learning models and the proposed CNN-LSTM model.
Table 4. Experimental results of all three scenarios with two baseline deep learning models and the proposed CNN-LSTM model.
ScenarioModelEvaluation Metrics
AccuracyPrecisionRecallF-measure
CNN89.6089.6089.6089.60
Acc.LSTM87.8088.0087.8087.90
CNN-LSTM95.2095.2095.2095.20
CNN86.4086.9086.4086.50
Gyro.LSTM84.1084.3084.1084.00
CNN-LSTM95.4095.4095.4095.40
CNN93.1093.1093.1093.10
Acc. + Gyro.LSTM89.6089.7089.6089.60
CNN-LSTM96.2096.3096.2096.30
Table 5. Confusion Matrix for the proposed CNN-LSTM.
Table 5. Confusion Matrix for the proposed CNN-LSTM.
WalkingJoggingStairsSittingStandingTypingTeethSoupChipsPastaDrinkingSandwichKickingCatchDribblingWritingClappingFoldingRecall
walking9521410000000063000198.35%
jogging2933500000000002210098.73%
stairs1197730000010062000797.89%
sitting004956161013251400070294.56%
standing200269534004401332510093.71%
typing00071891303200000240095.70%
teeth00051210171000010001198.83%
soup00033139207193400040294.94%
chips0008320109181682800000292.26%
pasta00050006697731300004196.26%
drinking00011151014158984000070090.52%
sandwich00195002592792160000193.41%
kicking01502000014696024000395.43%
catch0001100001102986510198.70%
dribbling0000000000000896360098.57%
writing000710002002021310163097.13%
clapping0000000000000003928399.36%
folding0011000000010423392898.41%
Precision99.50%99.68%97.99%91.66%95.20%98.78%99.41%97.77%96.43%93.31%94.63%88.39%97.56%95.45%97.27%94.69%98.83%97.48%96.20%
Table 6. Experimental results of stratified ten-fold cross validation.
Table 6. Experimental results of stratified ten-fold cross validation.
ModelEvaluation Metrics
Accuracy (SD)Precision (SD)Recall (SD)F-Measure (SD)
CNN90.92% (+/−0.53%)97.12% (+/−0.73%)97.17% (+/−1.41%)97.14% (+/−0.72%)
LSTM93.96% (+/−0.46%)99.48% (+/−0.51%)98.95% (+/−0.88%)99.21% (+/− 0.41%)
CNN-LSTM97.40% (+/−0.38%)98.60% (+/−2.59%)99.48% (+/−0.51%)99.02% (+/−1.30%)
Table 7. Performance (% accuracy) of comparison results.
Table 7. Performance (% accuracy) of comparison results.
ActivityPrevious WorkProposed CNN-LTSM
Acc.Gyro.Acc. + Gyro.Acc.Gyro.Acc. + Gyro.
Non-hand-orientedWalking87.7085.6089.1097.7096.6098.30
Jogging96.9093.6097.3099.4098.7098.70
Stairs85.5070.4084.0096.2096.5097.90
Sitting87.3062.8084.0092.2091.8094.60
Standing90.7059.0089.7095.3092.7093.70
Kicking82.9072.7084.4096.8096.9095.40
Average88.5074.0288.0895.2795.5396.43
Hand-oriented generalDribbling91.2090.6096.1095.8097.9098.60
Catch90.5088.7094.4098.2098.1098.70
Typing94.1083.3092.9095.7098.6095.70
Writing89.9077.6091.2097.5095.4097.10
Clapping95.0092.7096.6099.7099.3099.40
Teeth91.9081.6094.8098.6098.9098.80
Folding89.8085.3095.2094.6097.9098.40
Average91.7785.6994.4697.1698.0198.10
Hand-oriented eatingSoup86.6069.1087.3097.3091.0094.90
Chips78.8060.6080.0088.9094.3092.30
Pasta83.3068.3084.1093.7094.8096.30
Drinking80.9065.2080.8087.3089.3090.50
Sandwich72.7050.5070.9089.2088.6093.40
Average80.4662.7480.6291.2891.6093.48

Share and Cite

MDPI and ACS Style

Mekruksavanich, S.; Jitpattanakul, A.; Youplao, P.; Yupapin, P. Enhanced Hand-Oriented Activity Recognition Based on Smartwatch Sensor Data Using LSTMs. Symmetry 2020, 12, 1570. https://0-doi-org.brum.beds.ac.uk/10.3390/sym12091570

AMA Style

Mekruksavanich S, Jitpattanakul A, Youplao P, Yupapin P. Enhanced Hand-Oriented Activity Recognition Based on Smartwatch Sensor Data Using LSTMs. Symmetry. 2020; 12(9):1570. https://0-doi-org.brum.beds.ac.uk/10.3390/sym12091570

Chicago/Turabian Style

Mekruksavanich, Sakorn, Anuchit Jitpattanakul, Phichai Youplao, and Preecha Yupapin. 2020. "Enhanced Hand-Oriented Activity Recognition Based on Smartwatch Sensor Data Using LSTMs" Symmetry 12, no. 9: 1570. https://0-doi-org.brum.beds.ac.uk/10.3390/sym12091570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop