Next Article in Journal
Design of Surface Enhanced Raman Scattering (SERS) Nanosensor Array
Previous Article in Journal
ZnO Metal Oxide Semiconductor in Surface Acoustic Wave Sensors: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pattern Recognition of Cognitive Load Using EEG and ECG Signals

1
School of Electronic and Information Engineering, Southwest University, Chongqing 400715, China
2
Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Chongqing 400715, China
*
Author to whom correspondence should be addressed.
Submission received: 19 August 2020 / Revised: 3 September 2020 / Accepted: 6 September 2020 / Published: 8 September 2020
(This article belongs to the Section Wearables)

Abstract

:
The matching of cognitive load and working memory is the key for effective learning, and cognitive effort in the learning process has nervous responses which can be quantified in various physiological parameters. Therefore, it is meaningful to explore automatic cognitive load pattern recognition by using physiological measures. Firstly, this work extracted 33 commonly used physiological features to quantify autonomic and central nervous activities. Secondly, we selected a critical feature subset for cognitive load recognition by sequential backward selection and particle swarm optimization algorithms. Finally, pattern recognition models of cognitive load conditions were constructed by a performance comparison of several classifiers. We grouped the samples in an open dataset to form two binary classification problems: (1) cognitive load state vs. baseline state; (2) cognitive load mismatching state vs. cognitive load matching state. The decision tree classifier obtained 96.3% accuracy for the cognitive load vs. baseline classification, and the support vector machine obtained 97.2% accuracy for the cognitive load mismatching vs. cognitive load matching classification. The cognitive load and baseline states are distinguishable in the level of active state of mind and three activity features of the autonomic nervous system. The cognitive load mismatching and matching states are distinguishable in the level of active state of mind and two activity features of the autonomic nervous system.

1. Introduction

The aim of effective instructional design is to help learners construct or automate knowledge schemas by means of specific strategies at certain learning circumstances [1,2,3]. For example, collaborative group study helps learners with task-specific knowledge get better learning outcomes than the novices [1]; incorporating positive emotional design principles into multimedia lessons helps learners perform better on a subsequent retention test [2]; instructional animations are superior to static graphics in short learning sections, but not in long learning sections [3]. The key for effective learning is the matching of cognitive load and working memory [4]. Otherwise, when cognitive load exceeds working memory capacity, cognitive overload leads to bad learning outcomes [4]. Therefore, it is meaningful to monitor cognitive load in the learning process. An experienced teacher can judge cognitive load through an observation of the learner’s behavior and learning outcome [3]. However, teachers are often absent in e-learning, urging people to explore automatic cognitive load detection methods by using physiological measures.
Previous researchers have found that skin conductance response, which is controlled by the sympathetic nervous system (SNS), is significantly associated with cognitive load condition [5,6]. Besides, heart rate variability (HRV), controlled by both branches of the autonomic nervous system (ANS), has shown sensitivity to cognitive task load, conditions of event rate and task duration [7]. In addition to autonomic nervous measures, central nervous indices, e.g., theta power and alpha suppression of the electroencephalography (EEG), are also valid objective measures of average cognitive load [8,9,10]. These findings support that cognitive effort in the learning process has nervous responses encoded in the variation of quantitative parameters of several physiological signals.
Many machine learning methods have been applied to cognitive load detection, as shown in Table 1 [11,12,13,14,15,16,17,18,19,20,21,22,23,24,25]. The main idea of these methods is as follows. Firstly, define two or three categories of cognitive load conditions. Secondly, extract a set of physiological parameters as the features of cognitive load. Thirdly, train certain classifiers with data samples acquired from groups of subjects. Finally, obtain computational models for the pattern recognition of cognitive load through performance comparison of the classifiers. The strengths of the computational models are usually determined by the amount of data samples and subjects, the dimension of feature sets, the accuracy of the classifiers, and the validation methods of the models. For a given data sample set, a lower dimension of the feature set and a higher accuracy of the classifier and subject-independent validation method lead to a better pattern recognition model. As shown in Table 1, the subject-independent accuracy of the cognitive load recognition models still needs to be improved by means of more effective features, classifiers, or physiological signals.
In order to accurately recognize cognitive load conditions, the current work explored a large set of initial features of HRV and EEG, and selected a low-dimension critical feature subset by using sequential backward selection (SBS) and particle swarm optimization (PSO) algorithms. We trained several common classifiers with selected feature subset, and finally constructed effective pattern recognition models of cognitive load through subject-independent validation. Although previous work has shown that no explicit HRV parameters were continuously correlated with EEG parameters [26], we hypothesize that HRV and EEG parameters are complementary with each other, and their combination can improve the recognition of cognitive load conditions.

2. Materials and Methods

The dataset used in this work is available on Physiobank, and has been contributed by Igor Zyma, Sergii Tukaev, and Ivan Seleznov [27,28]. Firstly, necessary information about this dataset, e.g., subjects and mental task procedure, is introduced in this section. Secondly, we regrouped the data to form two binary classification problems of cognitive load. Thirdly, HRV and EEG parameters commonly used in literature were extracted as physiological features of cognitive load conditions. Lastly, we introduced the methods of feature selection and the criteria of classification accuracy.

2.1. Subjects

In total, 66 healthy right-handed volunteers (47 women and 19 men) were initially involved in the study. All subjects were 1st–3rd year students of the Taras Shevchenko National University of Kyiv, aged 18 to 26 years (18.6 ± 0.87 years). The subjects were eligible to enroll in the study if they had normal or corrected-to-normal visual acuity, normal color vision, and had no clinical manifestations of mental or cognitive impairment or verbal or non-verbal learning disabilities. Exclusion criteria were the use of psychoactive medication, drug or, alcohol addiction, and psychiatric or neurological complaints.

2.2. Experiment Procedure

The mental arithmetic task for inducing cognitive load was to continuously subtract the two-digit number from the four-digit number, and the accurate calculation times of each subject in four minutes were counted out after the arithmetic task. The subjects were rated according to the number of accurate calculations, and they were divided into two groups [27,28]. Group “G” had 24 subjects who performed good quality count (number of accurate calculations in 4 min: 21 ± 7.4); Group “B” had 12 subjects who performed bad quality count (number of accurate calculations in 4 min: 7 ± 3.6) [27,28]. Before the 4-min mental arithmetic task began, there were 3 min of baseline state recorded from each subject. ECG and EEG data were acquired throughout the experiment, but only the 3-min baseline data and the data in the first-minute calculation were kept in the dataset [27,28].
The EEGs were recorded monopolarly using the Neurocom EEG 23-channel system. The silver/silver chloride electrodes were placed on the scalp according to the International 10/20 scheme. All electrodes were referenced to the interconnected ear reference electrodes. Noise and artifacts in EEG data were removed by a 0.5 Hz cut-off frequency high-pass filter, a 45 Hz low-pass filter, a 50 Hz power-line notch filter and an independent component analysis (ICA) method [27,28].
We applied an adaptive threshold method based on wavelet decomposition to locate the position of R peaks in the ECG data and to obtain the RR interval series [29]. Due to the variation of heartbeat speed of different subjects, the number of RR intervals in one minute varies from 60 to 120. In order to obtain a compromise between the length of RR interval series and the number of subjects included in the dataset, we set 80 as the length of the RR interval series, and those without 80 RR intervals in one minute were excluded from the original dataset, keeping the data of 29 subjects (21 women and 8 men) for further analysis. The data selection and exclusion process is shown in Figure 1.

2.3. Grouping Rules

We aimed to solve two binary classification problems: cognitive load (CL) state vs. baseline (BL) state and cognitive load mismatching (CLMM) state vs. cognitive load matching (CLM) state. These two binary classification problems have practical values in e-learning. Distinguishing CL from BL helps the e-learning system to tell whether the learner is currently learning. Recognition of CLMM and CLM also helps the e-learning system to judge whether the learner is effectively learning. For the CL vs. BL classification problem, we divided the data into CL and BL groups: the CL group corresponding to data of the first-minute calculation, and the BL group corresponding to data of one-minute baseline. For the CLMM vs. CLM classification problem, we divided the data into CLMM and CLM groups: the CLMM group having first-minute calculation data of the subjects with bad quality count, and the CLM group having first-minute calculation data of the subjects with good quality count. We took effort to construct three different pattern recognition models for each of the above two binary classification problems, and these models were based on different options of physiological signals. The detailed information of the models is shown in Table 2.

2.4. Feature Extraction

In order to quantify the ANS activity, we extracted 27 linear and nonlinear parameters from the RR interval series as features of cognitive load conditions. These parameters were commonly applied to HRV analysis in literature [30,31,32,33]. Besides, six commonly used EEG parameters which measured the central nervous system (CNS) activity, e.g., powers of sub-band brainwaves [34], were also applied as features of cognitive load conditions. The HRV and EEG features are respectively shown in Table 3 and Table 4.

2.5. Balanced Sample Sets

As shown in Table 2, the sample sets of CLMM and CLM are unbalanced, which may cause the bias of the pattern recognition model to the majority samples. In order to avoid such bias, we adopted the Borderline-SMOTE1 algorithm to oversample the minority samples [35], so that the sample size of the CLMM group expanded to 18, the same to that of the CLM group.

2.6. Feature Selection and Classification

The original 27 HRV parameters and 6 EEG parameters were commonly applied to the analysis of ANS and CNS activities in literature [30,31,32,33,34]. However, they are not specific to the pattern recognition of cognitive load conditions. In terms of accuracy and computational efficiency, some features can even be a burden of the classification problems. Therefore, feature selection is necessary to find out the feature subset that is critical to the classification problems. The feature selection process is a combinatorial optimization problem to find a vector x * in the feature space X, so that
x X ,   f ( x * ) = min f ( x )
where x = [ x j ] 1 × M is a vector consisting of M features, with x j = 1 denoting the jth feature is selected, and x j = 0 denoting the jth feature is not selected. f ( ) is the evaluation function of the selected feature subset. In order to keep the complementary information among the features in the subset, we chose sequential backward selection (SBS) algorithm and particle swarm optimization (PSO) algorithm to perform feature selection [36,37]. SBS eliminates the least important feature from the subset at each iteration, until there is only one feature left in the subset. On the one hand, SBS is good at reducing the dimension of the feature subset. On the other hand, PSO has a strong ability of global optimization [38]. Therefore, when combining PSO and SBS, we firstly applied PSO to get an optimal feature subset which has the best evaluation function value in the searching process, and then used SBS to further reduce the dimension of the optimal feature subset. The quality of selected feature subset in each iteration of SBS and PSO process was evaluated by evaluation functions, i.e., f1 for SBS and f2 for PSO, which are calculated as in Equations (2) and (3):
f 1 = n = 1 N n correct n total
f 2 = α × # F e a t u r e s # A l l   F e a t u r e s + ( 1 α ) × E r r o r R a t e E R
where N is the number of subjects (N = 36 for CLMM vs. CLM, and N = 29 for CL vs. BL), n correct is the number of correctly classified samples of one subject, and n total is the number of samples of all subjects. α is the weighting coefficient, which is empirically set as α = 0.8. Moreover, #Features means the number of features currently selected, and #All Features is the total number of initial features. ErrorRate and ER respectively represent the mean error rates obtained by currently selected features and all features. n correct , n total , ErrorRate and ER were calculated in leave-one-subject-out cross validation. The leave-one-subject-out cross validation is to leave the samples belonging to one subject as the test set, and to repeat such test until the samples of every subject are tested. In the PSO feature selection process, the population size P = 30, the maximum iteration T = 50 in Model A and D and T = 100 in other models, the inertia weight ω = 0.7298, and the acceleration constants C 1 =   C 2 = 2 .
Table 4. EEG Features of Cognitive Load.
Table 4. EEG Features of Cognitive Load.
EEG IndexDescriptionRelation with CNS Activity
DPDelta band (1–4 Hz) powerA measure of unconscious mind [34].
TPTheta band (4.1–5.8 Hz) powerA measure of subconscious mind [34].
APAlpha band (5.9–7.4 Hz) powerA measure of relaxed mental state [34].
BPBeta band (13–19.9 Hz) powerA measure of active state of mind [34].
GPGamma band (20–25 Hz) powerA measure of hyper brain activity [34].
WEWavelet entropyA measure of energy distribution of EEG at different scales [39].
In this work, we chose DB4 as wavelet packet decomposition function with the scale of 7 to calculate the WE.
Four kinds of classifiers, namely support vector machine (SVM, kernel function types: ‘quadratic’ and ‘rbf’), K-nearest neighbor (KNN) and decision tree (DT), were adopted to solve the above-mentioned binary classification problems, and they were trained by the samples of N 1 subjects. Sensitivity (Sens.), specificity (Spec.), precision (Prec.), accuracy (Acc.), F1-score (F1) and the area under receiver operating characteristic (AUC) of the leave-one-subject-out cross validation were calculated to evaluate the performance of each classifier, as shown in Equations (4)–(9):
Sens = T P T P + F N
Spec = T N T N + F P
Prec = T P T P + F P
Acc = T P + T N T P + F P + T N + F N
F 1 = 2 S e n s S p e c ( S e n s + S p e c )
AUC = 1 2 ( T P T P + F N + T N T N + F P )
For each of the binary classification problems, we set samples in the CL and CLM categories as the positive ones, and samples in the BL and CLMM categories as the negative ones. In Equations (4)–(9), TP denotes the number of correctly classified samples in the positive category, TN denotes the number of correctly classified samples in the negative category, FP is the number of incorrectly classified samples in the negative category, and FN is the number of incorrectly classified samples in the positive category.

3. Results

3.1. Parameter Settings of Entropy Features

The value of entropy index depends on the settings of embedding dimension m, tolerance threshold r and delay time τ . The delay time τ was calculated with the mutual information method proposed in literature [40]. For the embedding dimension m and tolerance threshold r, we firstly set an initial variation range of m and r, and then determined the values of m and r through Mann–Whitney U test. The appropriate values of m and r resulted in the entropy indices, which were significantly different between the positive group (CL or CLM) and the negative group (BL or CLMM). The detailed parameter settings are shown in Table 5.

3.2. Results of Feature Selection

Figure 2, Figure 3 and Figure 4 respectively show the number of selected features and corresponding mean accuracy in the process of feature selection. Although the evaluation function of PSO in Equation (3) takes feature dimension into consideration, the dimension of the best feature subset found by PSO is still high, e.g., 6 for KNN (CL vs. BL) in Figure 2a, 7 for SVM (rbf) (CLMM vs. CLM) in Figure 3a, and 9 for KNN (CLMM vs. CLM) in Figure 4a. Considering the relatively small sample amount in the dataset, the higher dimension of feature set leads to higher risk of over fitting.
Therefore, we made a compromise between the accuracy of the classifier and the dimension of selected features, i.e., decreasing the dimension of selected features at the cost of a small decrease of mean accuracy. As marked in Figure 2, Figure 3 and Figure 4, if we constrain the dimension of the feature subset to be no more than 4, the highest mean accuracies with HRV features, EEG features and HRV+EEG features, were respectively obtained by DT (CL vs. BL and CLMM vs. CLM) and SVM (rbf) (CLMM vs. CLM) in Figure 2, KNN (CLMM vs. CLM) and DT (CL vs. BL) in Figure 3, and SVM_quadratic (CLMM vs. CLM) and DT (CL vs. BL) in Figure 4. Compared with the SBS algorithm alone, PSO and SBS found better 3-dimension feature subsets for models B, D and E. As shown in Table 6, the cost of doing this is that the computation time increases more than ten times.
The performance indices of the classifiers in Models A–F are shown in Table 6. Compared with single-modal signal (HRV or EEG), the fusion of HRV and EEG obtained better pattern recognition performances for CL vs. BL and CLMM vs. CLM problems. In Models C and F, we can see that the features of EEG and those of HRV are complementary to each other, so that the combination of EEG and HRV features improves the distinguishability of the positive and negative samples, verifying the hypothesis made in the introduction section. The confusion matrices of Models C and F are shown in Table 7. The bold parts in Table 6 and Table 7 represent the best classification results for each model.

3.3. Validation with E-Learning Data

We acquired e-learning data from real-world math courses for model validation. Five subjects took our e-learning math course and had their EEG and ECG data recorded. As our EEG device had 128 channels, the sub-band brainwave energy was very different from that of the 23-channel EEG data in literature [27,28]. Therefore, we only used the ECG data to validate our Model A. Each subject provided 10 min of ECG data, including 5 min of baseline state and 5 min of e-learning state. After eliminating noisy data, the sample size of the validation data set is 41 (21 of CL state and 20 of BL state). Each sample was described as 3-dimension vector of the critical ECG features of Model A, i.e., Area, LF and ApEn. Finally, we got 65.5% F1 score of the validation accuracy of Model A in the real e-learning status, and the confusion matrix is shown in Table 8.

4. Discussion

Compared with the previous research in Table 1, our work not only considered the CL vs. BL classification problem, but also explored the pattern recognition of CLMM vs. CLM. For the CL vs. BL classification problem, the best subject-independent results of the previous research are those in literature [11], i.e., 90% of Acc, 86% of Sens. and 95% of Spec. with SD1, SD2 and a complex measure of HRV named En(0.2). The best subject-independent results of CL vs. BL classification in this work are given by the DT classifier of Model C in Table 6. As shown in Table 6, better performance indices than those in [11] were obtained, with three HRV features and one EEG feature. If we use the same dimension of features to that in [11], the SVM (quadratic) classifier with three features, namely AP_O1, GP_O1 and Mean, got the Acc., Sens. and Spec. indices of 90.7%, 96.3% and 85.2%, respectively. For the CLMM vs. CLM problem, we use two HRV features and two EEG features to get better classification results than those in literature [19]. By using the real e-learning ECG data, we got a validation accuracy of 65.5% F1 score, much higher than that of a random guess.
It is worth noting that the relatively small amount of samples and subjects has limited further exploration of the generalization performance of the models proposed in this work. Although the validation accuracy of Model A is better than that of random guess, it is far away from the requirement of real application. There are two ways to improve the accuracy of CL vs. BL and CLMM vs. CLM classification. The first one is to enlarge the amount of data samples, and the second one is to use other math tasks which elicit CLMM and CLM states of the college students (i.e., the subjects). In our validation data acquisition, we found that subtracting a two-digit number from a four-digit number was too easy for the college students, and the CLMM state failed to be elicited.

5. Conclusions

By using the combination of one EEG feature (BP_F4) and three HRV features (Mean, LF and ApEn), the DT classifier has classified the CL and the BL states with the accuracy of 96.3%, showing that the CL and BL states are distinguishable in the level of active state of mind, the average level of ANS activity, the combined activities of SNS and PNS, and the competition between SNS and PNS. For the classification of CLMM and CLM states, the SVM (quadratic) classifier, two EEG features (BP_T4, BP_O1) and two HRV features (MFD, TFC) have obtained the accuracy of 97.2%, showing that the CLMM and CLM states are distinguishable in the level of active state of mind, the total and the average fluctuation of ANS activity. The current work has application value in practice, because it provides an objective quantitative method for the monitoring of cognitive load conditions in e-learning.

Author Contributions

Conceptualization, R.X. and W.W.; Methodology, R.X.; Software, F.K.; Validation, R.X., F.K. and X.Y.; Formal Analysis, R.X.; Investigation, F.K.; Resources, X.Y.; Data Curation, R.X.; Writing—Original Draft Preparation, R.X.; Writing—Review & Editing, W.W.; Visualization, W.W.; Supervision, G.L.; Project Administration, G.L.; Funding Acquisition, W.W. and G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China (Grant No. 61,103,132 and No. 61,872,301).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zambrano, J.; Kirschner, F.; Sweller, J.; Kirschner, P.A. Effects of prior knowledge on collaborative and individual learning. Learn. Instr. 2019, 63, 101214. [Google Scholar] [CrossRef]
  2. Le, Y.; Liu, J.; Deng, C.; Dai, D.Y. Heart rate variability reflects the effects of emotional design principle on mental effort in multimedia learning. Comput. Hum. Behav. 2018, 89, 40–47. [Google Scholar] [CrossRef]
  3. Wong, A.; Leahy, W.; Marcus, N.; Sweller, J. Cognitive load theory, the transient information effect and e-learning. Learn. Instr. 2012, 22, 449–457. [Google Scholar] [CrossRef]
  4. Sweller, J. Cognitive Load Theory. Psychol. Learn. Motiv. Cogn. Educ. 2011, 55, 37–76. [Google Scholar]
  5. Johannessen, E.; Szulewski, A.; Radulovic, N.; White, M.; Braund, H.; Howes, D.; Rodenburg, D.; Davies, C. Psychophysiologic measures of cognitive load in physician team leaders during trauma resuscitation. Comput. Hum. Behav. 2020, 111, 106393. [Google Scholar] [CrossRef]
  6. MacPherson, M.K.; Abur, D.; Stepp, C.E. Acoustic Measures of Voice and Physiologic Measures of Autonomic Arousal during Speech as a Function of Cognitive Load. J. Voice 2017, 31, 504.e1–504.e9. [Google Scholar] [CrossRef]
  7. Hughes, A.M.; Hancock, G.M.; Marlow, S.L.; Stowers, K.; Salas, E. Cardiac Measures of Cognitive Workload: A Meta-Analysis. Hum. Factors 2019, 61, 393–414. [Google Scholar] [CrossRef]
  8. Kardan, O.; Adam, K.C.; Mance, I.; Churchill, N.W.; Vogel, E.K.; Berman, M.G. Distinguishing cognitive effort and working memory load using scale-invariance and alpha suppression in EEG. NeuroImage 2020, 211, 116622. [Google Scholar] [CrossRef]
  9. Castro-Meneses, L.J.; Kruger, J.-L.; Doherty, S. Validating theta power as an objective measure of cognitive load in educational video. Educ. Technol. Res. Dev. 2019, 68, 181–202. [Google Scholar] [CrossRef]
  10. Jimenez-Guarneros, M.; Gomez-Gil, P. Custom Domain Adaptation: A New Method for Cross-Subject, EEG-Based Cognitive Load Recognition. IEEE Signal Process. Lett. 2020, 27, 750–754. [Google Scholar] [CrossRef]
  11. Hasanbasic, A.; Spahic, M.; Bosnjic, D.; Adzic, H.H.; Mesic, V.; Jahic, O. Recognition of stress levels among students with wearable sensors. In Proceedings of the 2019 18th International Symposium INFOTEH-JAHORINA (INFOTEH), East Sarajevo, Bosnia and Herzegovina, 20–22 March 2019; pp. 1–4. [Google Scholar] [CrossRef]
  12. Melillo, P.; Bracale, U.; Pecchia, L. Nonlinear Heart Rate Variability features for real-life stress detection. Case study: Students under stress due to university examination. Biomed. Eng. Online 2011, 10, 96. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Cheema, A.; Singh, M. An application of phonocardiography signals for psychological stress detection using non-linear entropy based features in empirical mode decomposition domain. Appl. Soft Comput. 2019, 77, 24–33. [Google Scholar] [CrossRef]
  14. Wang, Q.; Sourina, O. Real-Time Mental Arithmetic Task Recognition from EEG Signals. IEEE Trans. Neural Syst. Rehabil. Eng. 2013, 21, 225–232. [Google Scholar] [CrossRef] [PubMed]
  15. Al-Shargie, F.M.; Kiguchi, M.; Badruddin, N.; Dass, S.C.; Hani, A.F.M.; Tang, T.B. Mental stress assessment using simultaneous measurement of EEG and fNIRS. Biomed. Opt. Express 2016, 7, 3882–3898. [Google Scholar] [CrossRef] [Green Version]
  16. McDuff, D.J.; Hernandez, J.; Gontarek, S.; Picard, R.W. Cogcam: Contact-free measurement of cognitive stress during computer tasks with a digital camera. Presented at the in CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016. [Google Scholar]
  17. Ahn, J.W.; Ku, Y.; Kim, H.C. A Novel Wearable EEG and ECG Recording System for Stress Assessment. Sensors 2019, 19, 1991. [Google Scholar] [CrossRef] [Green Version]
  18. Xia, L.; Malik, A.S.; Subhani, A.R. A physiological signal-based method for early mental-stress detection. Biomed. Signal Process. Control 2018, 46, 18–32. [Google Scholar] [CrossRef]
  19. Dimitrakopoulos, G.N.; Kakkos, I.; Dai, Z.; Lim, J.; deSouza, J.J.; Bezerianos, A.; Sun, Y. Task-Independent Mental Workload Classification Based Upon Common Multiband EEG Cortical Connectivity. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1940–1949. [Google Scholar] [CrossRef]
  20. Yu, J.; Liu, G.Y.; Wen, W.H.; Chen, C.W. Evaluating cognitive task result through heart rate pattern analysis. Healthc. Technol. Lett. 2020, 7, 41–44. [Google Scholar] [CrossRef]
  21. Wang, C.; Guo, J. A data-driven framework for learners’ cognitive load detection using ECG-PPG physiological feature fusion and XGBoost classification. Procedia Comput. Sci. 2019, 147, 338–348. [Google Scholar] [CrossRef]
  22. Das Chakladar, D.; Dey, S.; Roy, P.P.; Dogra, D.P. EEG-based mental workload estimation using deep BLSTM-LSTM network and evolutionary algorithm. Biomed. Signal Process. Control 2020, 60, 101989. [Google Scholar] [CrossRef]
  23. Barua, S.; Ahmed, M.U.; Begum, S. Towards Intelligent Data Analytics: A Case Study in Driver Cognitive Load Classification. Brain Sci. 2020, 10, 526. [Google Scholar] [CrossRef] [PubMed]
  24. Plechawska-Wójcik, M.; Tokovarov, M.; Kaczorowska, M.; Zapała, D. A Three-Class Classification of Cognitive Workload Based on EEG Spectral Data. Appl. Sci. 2019, 9, 5340. [Google Scholar] [CrossRef] [Green Version]
  25. Fan, X.; Zhao, C.; Zhang, X.; Luo, H.; Zhang, W. Assessment of mental workload based on multi-physiological signals. Technol. Health Care 2020, 28, S67–S80. [Google Scholar] [CrossRef] [PubMed]
  26. Hillmert, M.; Bergmuller, A.; Minow, A.; Raggatz, J.; Bockelmann, I. Psychophysiological strain correlates during cognitive workload A laboratory study using EEG and ECG. Zent. Fur Arb. Arb. Und Ergon. 2020, 70, 149–163. [Google Scholar]
  27. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
  28. Zyma, I.; Tukaiev, S.; Seleznov, I.; Kiyono, K.; Popov, A.; Chernykh, M.; Shpenkov, O. Electroencephalograms during Mental Arithmetic Task Performance. Data 2019, 4, 14. [Google Scholar] [CrossRef] [Green Version]
  29. Wen, W.H.; Liu, G.Y.; Mao, Z.H.; Huang, W.J.; Zhang, X.; Hu, H.; Yang, J.; Jia, W. Toward Constructing a Real-time Social Anxiety Evaluation System: Exploring Effective Heart Rate Features. IEEE Trans. Affect. Comput. 2020, 11, 100–110. [Google Scholar] [CrossRef]
  30. Carr, O.; de Vos, M.; Saunders, K.E.A. Heart rate variability in bipolar disorder and borderline personality disorder: A clinical review. Évid. Based Ment. Health 2018, 21, 23–30. [Google Scholar] [CrossRef]
  31. Zhu, J.; Ji, L.; Liu, C. Heart rate variability monitoring for emotion and disorders of emotion. Physiol. Meas. 2019, 40, 064004. [Google Scholar] [CrossRef]
  32. Hua, Z.; Chen, C.; Zhang, R.; Liu, G.; Wen, W. Diagnosing Various Severity Levels of Congestive Heart Failure Based on Long-Term HRV Signal. Appl. Sci. 2019, 9, 2544. [Google Scholar] [CrossRef] [Green Version]
  33. Xie, J.; Wen, W.; Liu, G.; Li, Y. Intelligent Biological Alarm Clock for Monitoring Autonomic Nervous Recovery During Nap. Int. J. Comput. Intell. Syst. 2019, 12, 453–459. [Google Scholar] [CrossRef] [Green Version]
  34. Alarcao, S.M.; Fonseca, M. Emotions Recognition Using EEG Signals: A Survey. IEEE Trans. Affect. Comput. 2019, 10, 374–393. [Google Scholar] [CrossRef]
  35. Han, H.; Wang, W.-Y.; Mao, B.-H. Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning. In International Conference on Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  36. Zhang, J.; Yin, Z.; Cheng, P.; Nichele, S. Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review. Inf. Fusion 2020, 59, 103–126. [Google Scholar] [CrossRef]
  37. Kennedy, J.; Eberhart, R.C. Particle swarm optimization. In International Conference on Networks; IEEE: Piscataway, NJ, USA, 1995; pp. 1942–1948. [Google Scholar]
  38. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  39. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  40. Chen, Y.; Yang, H. Multiscale recurrence analysis of long-term nonlinear and nonstationary time series. Chaos Solitons Fractals 2012, 45, 978–987. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the data selection and exclusion.
Figure 1. The flowchart of the data selection and exclusion.
Sensors 20 05122 g001
Figure 2. The results of HRV feature selection for Models A and D. (a) SBS and PSO, and (b) SBS.
Figure 2. The results of HRV feature selection for Models A and D. (a) SBS and PSO, and (b) SBS.
Sensors 20 05122 g002
Figure 3. The results of EEG feature selection for Models B and E. (a) SBS and PSO, and (b) SBS.
Figure 3. The results of EEG feature selection for Models B and E. (a) SBS and PSO, and (b) SBS.
Sensors 20 05122 g003
Figure 4. The results of EEG+HRV feature selection for Models C and F. (a) SBS and PSO, and (b) SBS.
Figure 4. The results of EEG+HRV feature selection for Models C and F. (a) SBS and PSO, and (b) SBS.
Sensors 20 05122 g004
Table 1. Related work in cognitive load recognition.
Table 1. Related work in cognitive load recognition.
Study# Subjects# Features# CategoriesClassifierSignalsBest AccuracyValidation Approach
Hasanbasic [11]10123SVMECG, EDA91.00%SD
Melillo [12]4232LDAECG90.00%SI
Cheema [13]3052LS-SVMPCG96.67%SD
Wang [14]10322PCA, SVMEEG97.14%SD
Al-Shargie [15]2292SVMEEG, fNIRS95.10%SD
McDuff [16]1072Naïve BayesPPG (HR, HRV, BR)86.00%SI
Ahn [17]1442SVMECG, EEG87.50%SD
Xia [18]2242PCA, SVMEEG, ECG79.54%SD
Dimitrakopoulos [19]28232SVMEEG86.00%SD
Yu [20]2042ELMECG84.75%SI
Wang [21]160-2LFDM, XGBoostECG, PPG97.2%-
Das Chakladar [22]4862BLSTM-LSTMEEG86.33%-
Barua [23]66422Random ForestHRV, GSR, RESP78.00%SD
Plechawska [24]11523KNNEEG91.50%SI
Fan [25]2053SVM, PCAEEG, ECG80.00%SI
SD: subject-dependent. In this case, samples belonging to one subject appear both in the training dataset and in the validation dataset, usually causing over-optimistic accuracy of the classifier. SI: subject-independent, which means subjects and samples belonging to the validation dataset are totally new to the trained classifier. LS-SVM: least-square support vector machine; PCG: phonocardiography; PPG: photoplethysmography; ELM: extreme learning machine; LFDM: linear feature dependency modeling; XGBoost: eXtreme Gradient Boosting; BLSTM-LSTM: a combination of bidirectional long short-term memory (BLSTM) and long short-term memory (LSTM) networks. PCG: phonocardiography; PPG: photoplethysmography; EDA: electrodermal activity; fNIRS: functional near-infrared spectroscopy; HR: heart rate; BR: breathing rate; RESP: respiration. #: number.
Table 2. Grouping rules.
Table 2. Grouping rules.
Group# Subjects# Male# FemaleModel NamePhysiological Signal
CL vs. BL27 vs. 278 vs. 719 vs. 20Model AHRV
Model BEEG
Model CHRV and EEG
CLMM vs. CLM9 vs. 183 vs. 56 vs. 13Model DHRV
Model EEEG
Model FHRV and EEG
Among the data of 29 subjects, the BL data of two subjects and the CL data of another two subjects were heavily distorted by noise and eliminated from the dataset. #: number.
Table 3. HRV features of cognitive load.
Table 3. HRV features of cognitive load.
IndicesDescriptionRelation with ANS Activity
SDRRStandard deviation of RR intervalsA measure of HRV in time domain [26], which reflects the activities of SNS and PNS, mainly SNS activity [27].
RMSSDSquare root of the mean squared differences of successive RR intervalsA measure of HRV at one adjacent beat scale, which reflects the vagal activity [27].
MeanMean of RR intervalsA measure of the average level of ANS activity [26].
AreaSummation of RR intervalsA measure of total amount of ANS activity in time domain.
MFDMean of the first differences of RR intervalsA measure of HRV at one adjacent beat scale, which reflects the average fluctuation of ANS activity [25].
pNN20Proportion of differences between successive RR intervals longer than 20 msA measure of HRV in time domain, which reflects the fluctuation of ANS activity.
pNN10Proportion of differences between successive RR intervals longer than 10 msA measure of HRV in time domain, which reflects the fluctuation of ANS activity.
HRVCHeart rate variation coefficient, calculated by the ratio of SD to MeanA measure of normalized fluctuation of ANS activity.
VLFThe power of RR intervals between 0 Hz and 0.04 HzA measure of SNS activity [28].
LFThe power of RR intervals between 0.04 Hz and 0.15 HzA measure of combined activities of SNS and PNS [26,27].
HFThe power of RR intervals between 0.15 Hz and 0.4 HzA measure of PNS activity [26,27].
TOTPWRThe power of RR intervals between 0 Hz and 0.4 HzA measure of total amount of ANS activity in frequency domain [26].
HF/(LF+HF)The ratio of HF/(LF+HF)A measure of normalized PNS activity.
LF/(LF+HF)The ratio of LF/(LF+HF)A measure of normalized PNS+SNS activity [26].
LF/HFThe ratio of LF/HFA measure of the balance between SNS and ANS [27].
EntropyPeEn, ApEn, MFEn, SampEnMeasures of the complexity of RR interval series caused by competition between SNS and PNS [27].
DFA (α1, α2, α1/α2)Detrend fluctuation analysisMeasures of the fractal properties of RR interval series caused by competition between SNS and PNS [27].
TFCTotal fluctuation coefficientA measure of the fluctuation of ANS activity in scales 1~M [33]. We set M = 10 in the current work.
PP (SD1, SD2, SD1/SD2)Poincaré PlotMeasures of short-term and long-term HRV, which reflects the fluctuation of ANS activity [26,27].
RLHERange of the local Hurst exponentsA measure of the complexity of RR interval series, which is controlled by competition between SNS and PNS [25].
PeEn: permutation entropy; ApEn: approximate entropy; SampEn: sample entropy; MFEn: multiscale fuzzy measure entropy; PNS: parasympathetic nervous system; α1 and α2 were calculated with the small scale (4 ≤ n ≤ 16) and large scale (16 ≤ n ≤ 32), respectively.
Table 5. Settings of entropy parameters based on MANN–WHITNEY U test.
Table 5. Settings of entropy parameters based on MANN–WHITNEY U test.
FeatureGroupMean ± SDEmbedding DimensionTolerance ThresholdSig.Description
ApEnCL0.67 ± 0.16m = 2r = 0.4 × SDRR0.002m varies from 1 to 3, and r varies from 0.1 × SDRR to 0.9 × SDRR
BL0.78 ± 0.14
CLMM0.62 ± 0.12m = 2r = 0.6 × SDRR0.041
CLM0.47 ± 0.18
SampEnCL1.17 ± 0.33m = 1r = 0.30.01m varies from 1 to 3, and r varies from 0.1 to 0.9
BL1.39 ± 0.29
CLMM0.70 ± 0.16m = 2r = 0.60.03
CLM0.52 ± 0.21
PeEnCL0.59 ± 0.03m = 6-0.009m varies from 3 to 7, and τ is calculated by mutual information method
BL0.61 ± 0.02
CLMM0.97 ± 0.13m = 3-0.017
CLM0.93 ± 0.40
MFEnCL0.33 ± 0.18m = 1r = 0.1<0.001m varies from 1 to 3, and r varies from 0.1 to 0.9. The scale of CL vs. BL and CLMM vs. CLM are 5 and 2, respectively
BL0.50 ± 0.13
CLMM1.25 ± 0.16m = 3r = 0.20.015
CLM1.02 ± 0.27
Sig.: significance of Mann–Whitney U test.
Table 6. Critical feature subsets of models A–F and performance indices of the classifiers.
Table 6. Critical feature subsets of models A–F and performance indices of the classifiers.
ModelClassifierCritical Feature SubsetMfsTfs (min)F1Prec. (%)Sens. (%)Spec. (%)AUCAcc. (%)
Model ASVM_qArea, LF, HF/(LF+HF)SBS11.90.8783.392.681.50.8787.0
SVM_rRMSSD, LF, MFEnSBS10.40.8382.185.281.50.8383.3
KNNArea, LF, LF/HFSBS9.60.8681.396.377.80.8787.0
DTArea, LF, ApEnSBS10.60.9192.388.992.60.9190.7
Model BSVM_qAP_Pz, BP_F7, BP_O2SBS29.30.7271.474.170.40.7272.2
SVM_rDP_F8, AP_Fp1, BP_PzSBS25.80.7875.981.574.10.7877.8
KNNDP_T3, TP_F8, AP_O1PSO and SBS1455.30.7273.170.474.10.7272.2
DTAP_Fp2, AP_Pz, BP_O1PSO and SBS2234.70.8278.192.674.10.8383.3
Model CSVM_qAP_O1, AP_A2A1, GP_O1, MeanSBS167.80.9392.692.692.60.9392.6
SVM_rWE_P3, Area, LF, ApEnSBS198.50.9192.388.992.60.9190.7
KNNTP_O1, Mean, LF, ApEnPSO and SBS762.90.9086.796.385.20.9190.7
DTBP_F4, Mean, LF, ApEnSBS178.50.9693.110092.60.9696.3
Model DSVM_qMFD, SampEn, MFEn SBS7.90.8810077.81000.8988.9
SVM_rCVrr, SD1, SD1/SD2PSO and SBS515.80.9110083.31000.9291.7
KNNApEn, SD1, SD1/SD2PSO and SBS558.30.8593.377.894.40.8686.1
DTHF/(LF+HF), α2/α1, TFCPSO and SBS555.00.9294.188.994.40.9291.7
Model ESVM_qDP_T4, AP_Pz,PSO and SBS1207.00.8086.772.288.90.8180.6
SVM_rWE_F4, WE_F7SBS29.00.7873.994.466.70.8180.6
KNNDP_Cz, BP_F3PSO and SBS1141.80.9294.188.994.40.9291.7
DTDP_T6, GP_T4SBS35.20.8581.094.477.80.8686.1
Model FSVM_qBP_T4, BP_O1, MFD, TFCSBS114.00.9710094.41000.9797.2
SVM_rGP_Fz, MFD, SampEn, SD2SBS125.80.9494.494.494.40.9494.4
KNNGP_T4, MFD, PeEn, TFCSBS200.80.9410088.91000.9494.4
DTAP_T4, LF, TFC, SD1/SD2SBS149.30.9289.594.488.90.9291.7
SVM_q: quadratic polynomial as kernel function of SVM. KNN: k-nearest neighbors using Euclidean distance weighting. SVM_rbf: rbf as kernel function of SVM. Fp2, F7, Pz, T3, C3, A2, A1, F4, T4, T5, F8, C4, O1, O2, F3, Fp1, Cz, P4 and T6 represent various channels of EEG signal. Naming rules for EEG features: feature name _ channel name. Mfs: Method of feature selection. Tfs: Time of feature selection.
Table 7. The confusion matrices of Models C and F.
Table 7. The confusion matrices of Models C and F.
ModelClassifierMfsClassified asCLBLCLMMCLM
Model CSVM_qSBSCL92.6%7.4%--
BL7.4%92.6%--
SVM_rSBSCL88.9%11.1%--
BL7.4%92.6%--
KNNSBS and PSOCL96.3%3.7%--
BL14.8%85.2%--
DTSBSCL1000%--
BL7.4%92.6%--
Model FSVM_qSBSCLMM--100%0%
CLM--5.6%94.4%
SVM_rSBSCLMM--94.4%5.6%
CLM--5.6%94.4%
KNNSBSCLMM--100%0%
CLM--11.1%88.9%
DTSBSCLMM--88.9%11.1%
CLM--5.6%94.4%
Table 8. The confusion matrix of validation using e-learning data.
Table 8. The confusion matrix of validation using e-learning data.
ClassifierClassified asCLBL
DTCL55.0%45.0%
BL19.0%81.0%

Share and Cite

MDPI and ACS Style

Xiong, R.; Kong, F.; Yang, X.; Liu, G.; Wen, W. Pattern Recognition of Cognitive Load Using EEG and ECG Signals. Sensors 2020, 20, 5122. https://0-doi-org.brum.beds.ac.uk/10.3390/s20185122

AMA Style

Xiong R, Kong F, Yang X, Liu G, Wen W. Pattern Recognition of Cognitive Load Using EEG and ECG Signals. Sensors. 2020; 20(18):5122. https://0-doi-org.brum.beds.ac.uk/10.3390/s20185122

Chicago/Turabian Style

Xiong, Ronglong, Fanmeng Kong, Xuehong Yang, Guangyuan Liu, and Wanhui Wen. 2020. "Pattern Recognition of Cognitive Load Using EEG and ECG Signals" Sensors 20, no. 18: 5122. https://0-doi-org.brum.beds.ac.uk/10.3390/s20185122

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop