Next Article in Journal
Application of Phase-Reversal Fresnel Zone Plates for Improving The Elevation Resolution in Ultrasonic Testing with Phased Arrays
Next Article in Special Issue
Forcecardiography: A Novel Technique to Measure Heart Mechanical Vibrations onto the Chest Wall
Previous Article in Journal
High-Sensitivity, Large Dynamic Range Refractive Index Measurement Using an Optical Microfiber Coupler
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Real-Time Heartbeat Classification: Evaluation of Nonlinear Morphological Features and Voting Method

by
Rajesh N V P S Kandala
1,
Ravindra Dhuli
2,
Paweł Pławiak
3,4,*,
Ganesh R. Naik
5,
Hossein Moeinzadeh
5,
Gaetano D. Gargiulo
5,6,7 and
Suryanarayana Gunnam
8
1
Department of ECE, GVPCE (A), Visakhapatnam 530048, India
2
Department of ECE, VIT University, Andhra Pradesh 522237, India
3
Department of Information and Communications Technology, Faculty of Computer Science and Telecommunications, Cracow University of Technology, Warsaw 24 st., F-3, 31-155 Krakow, Poland
4
Institute of Theoretical and Applied Informatics, Polish Academy of Sciences, Bałtycka 5, 44-100 Gliwice, Poland
5
The MARCS Institute, Western Sydney University, Milperra, NSW 2214, Australia
6
Department of Electrical Engineering and Information Technology (DIETI), “Federico II” The University of Naples, 80100 Naples, Italy
7
School of Engineering at Western Sydney University, Penrith, NSW 2747, Australia
8
Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Submission received: 20 September 2019 / Revised: 10 November 2019 / Accepted: 15 November 2019 / Published: 21 November 2019

Abstract

:
Abnormal heart rhythms are one of the significant health concerns worldwide. The current state-of-the-art to recognize and classify abnormal heartbeats is manually performed by visual inspection by an expert practitioner. This is not just a tedious task; it is also error prone and, because it is performed, post-recordings may add unnecessary delay to the care. The real key to the fight to cardiac diseases is real-time detection that triggers prompt action. The biggest hurdle to real-time detection is represented by the rare occurrences of abnormal heartbeats and even more are some rare typologies that are not fully represented in signal datasets; the latter is what makes it difficult for doctors and algorithms to recognize them. This work presents an automated heartbeat classification based on nonlinear morphological features and a voting scheme suitable for rare heartbeat morphologies. Although the algorithm is designed and tested on a computer, it is intended ultimately to run on a portable i.e., field-programmable gate array (FPGA) devices. Our algorithm tested on Massachusetts Institute of Technology- Beth Israel Hospital(MIT-BIH) database as per Association for the Advancement of Medical Instrumentation(AAMI) recommendations. The simulation results show the superiority of the proposed method, especially in predicting minority groups: the fusion and unknown classes with 90.4% and 100%.

1. Introduction

1.1. Aim of the Work

Nowadays, mortality rates are increasing due to noncommunicable diseases (NCDs) over infectious diseases. Annually about 70% of deaths are because of the NCDs worldwide. As per the World Health Organization (WHO) report, cardiovascular diseases (CVDs) are the primary cause of death among other NCDs [1]. The effect of CVDs is more in low and middle-income countries. The report demonstrates that this impact will continue further. This alarming scenario influences not only the health perspective, but also the socio-economic advancement of the country. Therefore, the need for adequate diagnosis and treatment for NCDs, especially for CVDs, is highly essential. This situation demands advancements in healthcare technology.
Cardiac arrhythmias are one of the significant sources of CVDs. All arrhythmias may not be fatal, but some need immediate treatment for the patient to survive. Arrhythmias may occur owing to erratic electrical impulse conduction or formation in the heart. Electrocardiograms (ECGs) are an essential tool to study the electrical activity of the heart. ECG discloses any variations in the heartbeat pattern. Clinicians have to explore the longer duration ECG records in the diagnosis process. However, this manual examination is tiresome because of low amplitude and subtle variations in the ECG [2]. Hence, computer-aided diagnosis (CAD) helps clinicians remarkably. CAD-based heartbeat classification is a significant task before the arrhythmia recognition.

1.2. State-of-the-Art

If we look at the literature on heartbeat classification, it can be observed that the Massachusetts Institute of Technology-Beth Israel hospital (MIT-BIH) arrhythmia database [3] is the majority choice. The essential literature of heartbeat classification using this database can be categorized into two types based on the assessment process, viz., class-oriented and subject-oriented. The majority among them are class-oriented based works [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22].
In the class-oriented approach, from the 16 types of beats including the normal ones in the MIT database, a part or an entire collection of beats are preferred for classification. In [4], 17 types of heartbeats including normal and pacemaker are classified using the features based on various power spectrum density methods. Later, a novel genetic algorithm is used to identify the optimum features to enhance the classification process. Finally, these selected features are fed to the various standard machine learning algorithms. In [19], 13 types of heartbeats are classified using the combination of higher-order statistics (HOS) of the ECG and Hermite basis representation features using a support vector machine (SVM) classifier. In [18], six types of heartbeats are classified by using a local fractal dimension based nearest neighbor classifier. In [21], seven types of heartbeats are classified using gray relational analysis. In [20], Ye et al. designed a heartbeat classification algorithm using dynamic and morphological ECG features. For the morphological feature extraction process, the combination of wavelet transform and the dimensionality reduction technique, namely independent component analysis (ICA), is implemented on the heartbeats. R–R intervals are used as dynamic features. These features are then fed to SVM for classifying 16 types of heartbeats. In [8], a novel genetic ensemble of classifiers machine learning method is proposed. A new genetic training coupled with genetic optimization is used to classify 17 types of heartbeats. In [17], statistical and nonlinear features are derived from the modes obtained from the empirical mode decomposition (EMD) algorithm. Later, these features are provided to one-against-one SVM for classifying five types of heartbeats. In [22], ventricular extra systole or ectopic beats are recognized with the help of morphology matching, R–R intervals, and clustering algorithms. In [6], 17 types of EG beats are classified using hexadecimal local patterns claculated from wavelet sub-bands. In [7], five primary types of heartbeats are classified using ensemble empirical mode decomposition (EEMD) based features subjected to sequential minimal optimization-SVM (SMO-SVM). Besides, Neural networks plays a crucial role in biological signal analysis [23]. Recently, deep learning-based class-oriented schemes come into the picture. Deep learning techniques are a part of machine learning techniques implemented based on more hidden neural networks. In [9,10] these works, 17 types of heartbeats are classified using 1D-CNN and a novel 3-layer deep genetic ensemble of classifiers.
In the subject-oriented approach, the entire MIT-BIH database is subdivided into five groups of heartbeats according to the American National Standards Institute/Advancement of Medical Instrumentation (ANSI/AAMI) EC57:1998 standard. The list of these groups is non-ectopic (N), supraventricular ectopic (S), ventricular ectopic (V), fusion (F), and unknown (Q). Again, two strategies are observed for classifying these distinct groups: intra-patient and inter-patient schemes. The fundamental disagreement between these two strategies is the separation of training and testing datasets. Intra-patient scheme based methods are widely explored in the literature [24,25,26,27,28,29,30]. However, these approaches have less impact in real-time scenarios. Because, in real-time applications, an unknown subject that usually undergoes the testing will be foreign to the constructed model. Thus, the model has to be adequate to capture the inter-individual variations among the ECG. While designing the intra-patient based model, there might be a chance of having common subject information in both training and testing. To mitigate such an issue, De Chazal et al. [31] introduced an inter-patient scheme based heartbeat classification. Here, the overall MIT-BIH database is separated into two groups. One group is assigned to training, and the other one is for testing by ensuring that there is no similar subject data in both groups.
The advantage of the aforementioned computer-aided expert systems can be exploited only after developing real-time systems. In literature, in recent years, some of the field-programmable gate array (FPGA) based ECG signal analysis systems are implemented. In [32], an FPGA based heartbeat classification system is developed using the least-squares linear-phase finite impulse response filter and feed-forward neural network. In [33], three types of common arrhythmia beats, namely, premature ventricular contraction, ventricular fibrillation, and heart block beat along with normal beats, are classified using a real-time FPGA implementation. In [34], an intra-patient scheme based on arrhythmia classification is implemented in the FPGA system. However, most of the successful FPGA implemented systems are followed by an intra-patient scheme. Very few methods are developed in real-time systems based on inter-patient schemes [35]. However, still, these systems failed in detecting rare abnormal beats accurately. Hence, there is a need for developing a new expert system that can succeed in identifying rare heartbeats.

1.3. Contribution

In this paper, we presented an efficient inter-patient heartbeat classification algorithm. For any pattern recognition process, identifying an appropriate set of features and classifier is highly significant. From [36], it is noticeable that ECG is a non-stationary, non-Gaussian signal derived from nonlinear systems. Hence, we employed a decomposition method, namely improved complete ensemble empirical mode decomposition (ICEEMD) to obtain features from the ECG beats. This technique is capable of disclosing the implicit information lying in the ECG. Later, different nonlinear measures like entropies and HOS are determined from the modes obtained after ICEEMD. These measures will serve as features for proper discrimination of the heartbeat groups. The fundamental difficulty in processing these groups is the class imbalance. Here, a significant fraction of the heartbeats is non-ectopic. Hence, the results may be biased toward the majority group, which is undesirable. Therefore, to alleviate such an issue, we followed an algorithmic level approach. To achieve this, we employed a majority voting scheme based classification. It is a type of ensemble classification. The advantage of ensemble classification is that it can reduce both variance and bias. In this work, we used different combinations of classifiers, namely, naïve Bayes, linear, and quadratic discriminant functions, J48, and consolidated J48 classifiers for majority voting.
The rest of the paper is ordered as follows: the ECG data set, training, and testing data division of AAMI labeling, experimental details and theoretical background of the methodology are presented in Section 2. Section 3 presents the simulation results of the proposed method. The comparison with existing works, limitations, and future directions are presented in Section 4. The conclusions of the work are presented in Section 5.

2. Methods

The block diagram of the proposed method is illustrated in Figure 1. The methodology consists of three stages including pre-processing, feature extraction on training and testing data, and a classification model for evaluation. In this section, the database used and the theoretical background of the used techniques are discussed.

2.1. Database

The proposed method is examined using the MIT-BIH arrhythmia database. MIT-BIH is a standard database widely explored for arrhythmia classification. It comprises of Holter monitoring records from several male and female patients. Each record duration is 30 minutes, sampled at 360 Hz. The records consist of both normal and abnormal beats of 15 types.
The annotation files available in the database are obtained from the chart recordings recognized by the experts. This file describes the ‘R’ peak locations and the labeling of normal and abnormal beats. Based on the recommendations of AAMI, class-labeling was assigned for discriminating various heartbeat groups.

AAMI Class Labeling Recommendations

According to the ANSI/AAMI EC57:1998 standard, within the annotation files, beat labels are divided into five groups, namely, N, S, V, F, and Q based on the physiological origin of the beats. Here, the mainly N group consists of normal and bundle branch block beats. S and V groups consist of ectopic beats, originated above and below Atrio Ventricular (A–V) junction of the heart, respectively. The F group consists of the combination of ventricular and normal beats. Unclassifiable beats are placed in the Q group. According to [31], the total number of available heartbeats are divided into training for modeling and testing for evaluation. Details of the number of heartbeats utilized for this work are presented in Table 1.

2.2. Pre-Processing

Pre-processing is an initial step in any data processing systems. Raw ECG signals will inherently have some artifacts. These may occur due to instrumental noise (power line interference), a physiological signal disturbance (muscular movements), or the environment where the experiment takes place. These artifacts are undesirable and diminish significant features in the ECG. Therefore, to attenuate the effect of this noise, we perform denoising as one of the pre-processing steps. For this, we used a filtering routine proposed by [37] with minimal modification. This operation comprises the following:
  • Mean separation from the noisy ECG,
  • Moving average filter of order five,
  • High-pass filter with cut-off frequency 1 Hz (for baseline wander suppression),
  • Low pass Butter worth filter with cut-off frequency 45 Hz (To suppress any left out high-frequency noise).
We need individual heartbeats from the long-term ECG recording for heartbeat classification. We perform a segmentation process after denoising. In the segmentation process, the annotation chart records with ‘R’ peak locations are utilized. From the annotation file, it is observed that there are a lot of variations among R-peak positions time-to-time. The difference between the R-peak positions is dynamic. Hence, we applied a window of length 300 samples on ECG signal to obtain an ECG segment that covers the QRS complex which is an important epoch in the ECG. Our segmentation process retains other important epochs like P and T waves, unlike centered R-peak distribution segmentation methods.

2.3. Feature Extraction

Feature extraction has a critical role in heartbeat classification. A feature provides crucial information about a signal and facilitates better discrimination of classes. From [36], it is evident that ECG is a non-stationary signal stemming from a nonlinear system. Hence, exploration of ECG with nonlinear methods can improve the performance of a model since they extract subtle information lying in ECG. Therefore, in the feature extraction stage, initially, we perform ICEEMD on ECG segments to get intrinsic mode functions (IMFs). Later, entropy and higher-order cumulants are extracted from the selected modes. In this section, the techniques employed and their support in the methodology development are briefly discussed.

2.3.1. ICEEMD

The EMD decomposes a given signal in a full data-dependent approach by exploiting the local characteristics. However, EMD is limited by “mode-mixing’’ problem while analyzing the real data [38]. Therefore, some noise-assisted data analysis methods can provide a solution. Here, noise is added in a controlled manner for developing new extrema. Thus, the local mean is limited to that of the original version where extrema are generated. A few among these noise assisted methods are EEMD [39] and CEEMD [40]. Among these methods, CEEMD provides a better solution to the mode-mixing problem. However, CEEMD has some limitations:
(i)
Some residual can be present in the modes.
(ii)
During the initial decomposition stages, information may appear “late” with undesired modes, when it is compared to EEMD.
To address these issues, Colominas et al. [41] introduced a new noise aided adaptive data analysis method called ICEEMD. The mathematical details of the ICEEMD are given below [41].
Notation used in algorithm: E l ( . ) = l t h EMD mode, M ( . ) = local mean of the signal, < . > = averaging operator, w ( j ) = realization of white Gaussian noise with zero mean and unit variance and x = input signal.
The algorithm steps:
  • Compute the local means of J realizations x ( j ) = x + β 0 E 1 ( w ( j ) ) , j = 1 , 2 , , J using EMD, to obtain first residue r 1 = < M ( x ( j ) ) > .
  • At the first stage ( l = 1 ), compute the first IMF:
    C 1 = x r 1 .
  • For l = 2 , , L , calculate r l as
    r l = < M ( r l 1 + β l 1 E l ( w ( j ) ) > .
  • Calculate the l t h mode as
    C l = r l 1 r l .
  • Go to step 3 for next l
Here, β l = ϵ 0 σ ( r l ) is used to obtain the desired SNR at each stage. We choose ϵ 0 = 0.2 . The resultant IMFs provide significant underlying features of the ECG signal. The ICEEMD is a beneficial tool used for analyzing non-stationary signals originating from nonlinear systems such as bio-signals. The main advantage of ICEEMD is: avoiding the spurious modes and reducing the amount of noise in the mode patterns. Thus, the decomposed IMFs capture the morphology of the signal. Later, entropy and statistical measures are calculated from the first six modes of each ECG segment.

2.3.2. Entropy Measures

Entropy measures the uncertainty in a given data. It is often used in signal processing and pattern recognition applications [42]. A high value of entropy maps to higher uncertainty (or) unpredictability. Entropy yields useful information for analyzing non-stationary signals [43]. In this work, we calculated Shannon [44], log energy, and norm entropies [45]. The entropy E must be an additive cost function such that E ( 0 ) = 0 and
E ( s ) = i E ( s i ) ,
where s is the probability of the given signal and i represents one of the discrete states. Various entropies are defined below:
  • Shannon Entropy:
    E Shannon ( s ) = i s i 2 log ( s i 2 ) ,
  • log Energy Entropy:
    E log   energy ( s ) = i log ( s i 2 ) ,
    with the convention log ( 0 ) = 0 .
  • norm Entropy: The l p norm entropy with 1 p is defined as
    E norm ( s ) = i | ( s i p ) | = | | s | | p p .

2.3.3. HOS

HOS provides a meaningful measure for analyzing non-stationary signals originating from nonlinear systems [46,47]. HOS represents the deviation from Gaussianity and can provide useful information from the non-Gaussian nature of ECG signals. In our work, we utilized second, third, and fourth-order cumulants as HOS. The mathematical details of the HOS can be found in [48].
We construct a feature vector of size 36 × 1 for each heartbeat (6 features × 6 modes = 36). Later, the training feature set is fed to a classifier for building a model, and that model is evaluated using a testing set.

2.4. Voting Scheme

The final goal of machine learning is to get better-generalized performance. We come across a question “which learning algorithm or classifier is preferred over the other ?’’. According to a “No free launch theorem’’ [49], there is no precise answer to this. One algorithm fits or performs well for a set of training and testing data and may fail for another. The learning algorithm overall performance depends on the prior information, distribution of data, amount of training data, and some cost functions. The performance generalization depends on the bias and variance errors. Always, there will be a trade-off between bias and variance. Ensemble classifiers form a better choice, to improve the performance generalization by reducing bias and variance. Combining several classifiers for the final decision is called an ensemble classification or mixture-of-experts model or modular classification.
The primary motivation behind the classifier ensemble is improving the classification performance using the complementary information offered by various classifiers. Kittler et al. [50] developed a scheme for combining classifiers using voting based on a set of rules: min-rule, max-rule, product-rule, sum-rule, and median-rule. From our experiments, we preferred product rule which outperforms others.
Mathematical Framework: Consider a pattern recognition model where a pattern y is to be assigned with one of the m possible classes ω 1 , ω 2 , , ω m . Say there are R number of classifiers used for combining. Let us assume that each classifier possesses a different representation of measurement vector x i , i = 1 , 2 , , R .
The density function for each class ω k in the measurement space is p ( x i | ω k ) and the prior probability is P ( ω k ) . We assume that the models are mutually exclusive.
From the Bayesian framework, y is assigned to the class ω j having a maximum posterior probability out of ω k classes:
assign y ω j if P ( ω j | x 1 , x 2 , . . , x R ) = max k P ( ω k | x 1 , x 2 , . . , x R ) .
Rewriting the posterior probability P ( ω k | x 1 , x 2 , . . , x R ) using the Bayes theorem:
P ( ω k / x 1 , x 2 , . . , x R ) = p ( x 1 , x 2 , . . , x R | ω k ) P ( ω k ) P ( x 1 , x 2 , . . , x R ) .
Here, P ( x 1 , x 2 , . . , x R ) can be expressed in terms of conditional measurement distribution as
p ( x 1 , x 2 , . . , x R ) = j = 1 m p ( x 1 , x 2 , . . , x R | ω j ) P ( ω j ) .
Product Rule: p ( x 1 , x 2 , . . , x R | ω j ) represents the joint probability distribution of the measurements computed by the classifiers. Assuming that these representations are statistically independent, we can rewrite the joint probability distribution as
p ( x 1 , x 2 , . . , x R | ω k ) = i = 1 R p ( x i | ω k ) .
Based on p ( x i | ω k ) , the measurement process model for i t h representation is developed. Substituting Equation (10) and Equation (9) into Equation (8)
P ( ω k | x 1 , x 2 , . . , x R ) = P ( ω k ) i = 1 R p ( x i | ω k ) j = 1 m p ( ω j ) i = 1 R p ( x i | ω j )
and using Equation (11) in Equation (7), we obtain the decision rule
assign y ω j if P ( ω j ) i = 1 R p ( x i | ω j ) = max k = 1 n P ( ω k ) i = 1 R p ( x i | ω k ) .
Rewriting in terms of the posterior probabilities obtained from the respective learning algorithms,
assign y ω j if 1 P ( R 1 ) ( ω j ) i = 1 R p ( ω j | x i ) = max k = 1 n 1 P ( R 1 ) ( ω k ) i = 1 R p ( ω k | x i ) .
Equation (13) represents the likelihood decision rule obtained after combining the posterior probabilities generated by different classifiers using the product rule.
In this work, we used five different classifiers for ensembling using a voting scheme to enhance the performance of the system: naïve Bayes [51], linear and quadratic discriminant functions [52], J48 [53], and J48 consolidated classifiers [54]. A brief description of these classifiers is given below.
Naïve Bayes Classifier: It is a probability-based learning algorithm developed on the Bayesian framework. According to Bayes theorem, an unknown y is categorized into the one among the R classes, with high posteriori probability:
y ω k if   arg   max ω k ω P ( ω | y ) P ( y ) ,
where ω = ω 1 , ω 2 , . , ω R is a vector of R classes.
Naïve Bayes is a modified version of Bayes classifier, based on the assumption that the features in an unknown example vector are independent. Therefore, posteriori probability can be written as
P ( ω | y ) = P ( y | ω ) P ( ω ) = P ( y 1 , y 2 , , y m | ω ) = P ( y 1 | ω ) P ( y 2 | ω ) P ( y m | ω ) P ( ω ) .
Hence, Equation (14) can be modified as
y arg   max ω k P ( ω = ω k ) i P ( y i | ω = ω k ) .
With this final rule, the naïve Bayes classifier operates. The parameters used for the Naïve Bayes Classifier is given in Table 2.
In general, the naïve Bayes classifier assumes that the given features follow the normal distribution. In Table 2, use the Kernel Estimator parameter set to false to follow this assumption. Supervised discretization converts a specific range of attribute values to binary values. Here, the term supervised is coined because the class information of the training instances is used for discretization. However, this process is possible only when the class labels are nominal. The advantage of supervised discretization in naïve Bayes classifier is present in [55].
Linear and Quadratic Discriminant Analysis Based Classifiers: The approach of discriminant analysis is to derive a decision boundary or a discriminant function based on the linear combinations of features that best separate the given classes. The assumption made is: examples from different categories follow Gaussian distribution. For instance, the discrimination function for two-class problems based on Bayes theory can be written as
( y μ 1 ) T Σ 1 1 ( y μ 1 ) + ln | Σ 1 | ( y μ 2 ) T Σ 2 1 ( y μ 2 ) > T ,
where μ 1 , μ 2 are the mean vectors of class1 and class 2, Σ 1 , Σ 2 are the covariance matrices of class 1 and class 2 and T is the threshold value.
The above function without further assumptions represents the quadratic discriminate function. If the covariance matrices Σ 1 = Σ 2 = Σ , then the discriminant function simplifies to a dot product.
x . y > constant ,
where x = Σ 1 ( μ 1 μ 2 ) , constant = 1 2 ( T μ 1 T Σ μ 1 μ 2 Σ μ 2 T ) . This decision rule represents the classification based on linear discriminant.
The parameters used for linear discriminant analysis (LDA) and quadrature discriminant analysis (QDA) classifiers are given below in Table 3.
Ridge parameters in the discriminant analysis classifiers reduce the overfitting problem by penalizing the large quantity coefficients. In our work, we use the default values as given in Table 3.
J48 Classifier: Recently, decision tree-based algorithms have become popular in machine learning strategies. In practice, J48 is an execution of popular C 4.5 algorithms proposed by Quinlan [53]. According to this algorithm, the decision process involves the construction of a tree based on the feature splitting. The superiority of matching y to a class label ω k ω depends on the choice of feature splitting based on the value of information gain.
Information gain is measured with the help of difference entropy as the difference between the entropy of the central node to the sum of entropies of the leaf nodes. It measures how well a given feature splits the training data under its class label. A feature node having high information gain is preferred.
J48 Consolidated (J48-C) Classifier: It is a consolidated version of C 4.5 classifier. “J48 consolidated’’ is an implementation of a consolidated tree’s construction algorithm, proposed by Arbelaiz et al. [54] in WEKA. The basic idea is building a single tree using several subsamples. In each iteration, we will find a better feature using information gain content similar to J48. After finding the best feature split, all the subsamples are divided using the same features. More details can be found in [54].The parameters used for J48 and J48-C classifiers are given in below Table 4.
J48 and J48-C classifiers are decision tree classifiers in which tree splitting criteria play a significant role. The above-mentioned parameters determine the growth and direction of the tree structures that influence the final model accuracy. Sub-tree raising considers raising of a sub-tree when pruning is enabled. The minimum number of objects determines the number of instances per leaf. Minimum description length (MDL) correction is a statistical measure like information gain to identify the best split tree. The number of folds determines the data used for error reduce pruning; here, one fold is for pruning and the other folds for building the tree.
All the parameters are fixed based on the final results. All the details of the parameters can be found in WEKA 3.9 version [56].

3. Results

In this work, we are classifying the five classes: N, V, S, F, Q. The training set is constructed with the array of records as DS1 = [101, 106, 108, 109, 112, 114, 115, 116, 118, 119, 122, 124,201, 203, 205, 207, 208, 209, 215, 220, 223, 230], and DS2 = [100, 103, 105, 111, 113, 117, 121, 123, 200, 202, 210, 212, 213, 214, 219, 221, 222, 228, 231, 232, 233, 234]. Here, the numerical values represent the patient record number. Four records (102, 104, 107 and 217) having paced beats are exempted from both DS1 and DS2 data sets.
We start with scatter plots for justifying the choice of features in discriminating against the heartbeats. Individual performance of five classifiers naïve Bayes, LDA, QDA, J48, and J48 consolidated is presented, and analysis using a voting scheme with various combinations of these classifiers is considered. The performance is illustrated for each combination. We used the WEKA 3.9 version (University of Waikato, New Zealand) [56] for implementing the classification algorithms and scatter plots. Data pre-processing and feature extraction is implemented using MATLAB 2018a (Mathworks, MA, USA). All the experiments are carried out in Windows 8, 8 GB RAM, and 64-bit operating system.

The Performance Measures

An algorithm’s efficiency can be validated with appropriate performance measures. In this work, Sensitivity (SEN), False Positive Rate (FPR), Positive Predictive Value (PPV), and Overall Accuracy (OA) are used as performance measures to compare with the state-of-the-art methods, following the AAMI recommendations. The confusion matrix required for calculating these measures is given in Table 5. For V and S classes, the measures are calculated as per [31]. For remaining classes, we followed [57].
Performance measure from Table 5 can be calculated as follows:
The sum measures of row-wise and column-wise calculations are:
R N = N n + N v + N s + N f + N q ; C N = N n + V n + S n + F n + Q n ; R V = V n + V v + V s + V f + V q ; C V = N v + V v + S v + F v + Q v ; R S = S n + S v + S s + S f + S q ; C S = N s + V s + S s + F s + Q s ; R F = F n + F v + F s + F f + F q ; C F = N f + V f + S f + F f + Q f ; R Q = Q n + Q v + Q s + Q f + Q q ; C Q = N q + V q + S q + F q + Q q ; R = R N + R V + R S + R F + R Q ; C = C N + C V + C S + C F + C Q ; R = C .
The false-positive and false negative values for each class are defined as below:
F N N = R N N n ; F P N = C N N n ; F N V = R V V v ; F P V = C V ( V v + F v + Q v ) ; F N S = R S S s ; F P S = C S ( S s + Q s ) ; F N F = R F F f ; F P F = C N F f ; F N Q = R Q Q q ; F P Q = C N Q q .
The other useful measures, true positives, and negatives can be calculated for Classes N,V,S,F,Q:
T P N = N n ; T N N = R ( R N + C N N n ) ; T P N = V v ; T N V = R ( R V + C V V v ) ; T P S = S s ; T N S = R ( R S + C S S s ) ; T P F = F f ; T N F = R ( R F + C F F f ) ; T P Q = Q q ; T N Q = R ( R Q + C Q Q q ) .
The performance measures are given by
SEN = T P T P + F N ; FPR = F P T N + F P ; PPV = T P T P + F P ; OA = T P N + T P V + T P S + T P F + T P Q R ,
where T P = True   Positive , T N = True   Negative , F P = False   Positive , a n d F N = False   Negative .
We present the scatter plots with marginal histograms on the testing data set DS2, for features in the two-dimensional feature space. These scatter plots reveal how different features spread in feature space, thereby revealing the relationship between different heartbeat classes. Figure 2 shows the two-dimensional scatter plot between cumulant 2 of IMF1 to norm entropy value of IMF2. In this plot, we can observe that N and V beats are dominantly spread across space. In addition, the histogram plots also reveal the good discrimination between N, V, and Q classes out of the five classes. The next plot from Figure 3 gives the relation between the log energy entropy of IMF1 to cumulant 2 of IMF1. In this figure, we can see the spreading of N, V, S, and F classes in the space. In particular, this space provides good discrimination between N, V, and S classes. From Figure 4, Figure 5 and Figure 6, we can observe that log energy entropy values extracted from different IMFs provide a good perception of N, V, and S classes.
In the same way, Figure 7 and Figure 8 give better discrimination of Q beats, which are very rare indeed. In these figures, the characteristic feature is the norm entropy. In addition, different combinations of features with norm entropy reveal different class spreads and discrimination capabilities. As a whole, we can say that the combinations of selected features from different IMFs can predict the required hypothesis.
After dividing the training and testing feature sets, we need to learn a model for classification. In this work, we used an ensemble learner for classification. Ensemble classifiers use multiple learning algorithms and combine all the decisions. It can be more accurate than the individual classifiers. The main advantage of the ensemble classifiers is that we can achieve low bias error and low variance error. Ensembles using multiple trained (high variance/ low bias) models can average out of the variance, leaving just the bias. In addition, ensemble classifiers are preferred for imbalanced datasets. Our DS1 and DS2 datasets are highly imbalanced with majority N group class and minor F and Q classes. Therefore, in this work, we used a voting scheme based on product rule to ensemble the classifiers. The individual classifier performance on DS2 (testing data) is presented in Table 6 and Table 7. Confusion matrices calculated for LDA, QDA, naïve Bayes, J48 and J48-C classifiers are shown in Table 6. The performance measures for the corresponding matrices based on Table 5 are presented in Table 7.
From this table, we can see that each classifier yields different prediction. LDA and J48 give better classification for the N and V classes. It is an understandable phenomenon because of the dominating number of examples in N and V. LDA and J48-C provides better discrimination to Q group. The other classifiers J48-C and naïve Bayes are providing better SEN results for F group. Finally, S class is predicted accurately by J48-C and QDA classifiers. The other important point is, although all classifiers yield better results for the specific group of categories, the OA is dominated by the N class discrimination. Therefore, it is noticeable that OA is no longer a useful performance measure for imbalanced data classification.
In Table 8, the confusion matrix after combining J48, LDA, and naïve Bayes classifiers using the voting scheme is presented. The corresponding performance measures are demonstrated in Table 9. From the results, it is evident that this combination yields better results for N, V, F, and Q classes and average result for S class. The critical point is N, and S classes have more morphological similarities. Therefore, individual classifiers are giving complementary results for N and S. However, this ensemble selection enhances the prediction generalization for both classes.
Similarly, we performed ensemble voting for different combinations and the results are presented in Table 10, Table 11, Table 12, Table 13, Table 14 and Table 15. Each combination provides various enhanced results in some aspects.
As mentioned earlier, the dataset is dominated by N, V, and S classes, respectively. The F and Q classes are sporadic. Therefore, in some works, only N, V, and S classes are considered for classification. We provide the results for such schemes in Table 16, Table 17, Table 18, Table 19, Table 20, Table 21, Table 22, Table 23, Table 24 and Table 25. Here, first results are also presented for individual classifiers; later, the ensemble voting scheme is performed on different combinations of classifiers. Each one gives better classifications than individual classifiers.

4. Discussion

This section contains a discussion on simulating the proposed methodology illustrated in Figure 1. In this work, we employ an adaptive non-stationary and nonlinear decomposition method, namely ICEEMD, to analyze the ECG heartbeats. ICEEMD produces a local and entirely data-driven separation of a signal in the form of fast and slow oscillations called IMFs. The main advantage of ICEEMD is that it successfully avoids the spurious nodes and reduces the amount of noise in the modes.
Later, six nonlinear morphological features: higher-order cumulants, log, Shannon energy, and norm entropies are extracted from the first six IMFs of each heartbeat, to generate a 36 × 1 feature vector. Then, these feature vectors are divided based on training and testing sets DS1 and DS2 as specified above. Statistics (median and interquartile range) of these features for each class are presented in Table 26. Variation of attributes corresponding to different heartbeats can be observed from this table.
In Table 6, we presented the individual performance of various classifiers on the given problem. Here, all classifier models offer separate results for all the classes. Each model performs well for a specific class or classes. However, it fails in providing the overall better performance. For example, the S class contains 1837 beats, the J48, LDA, and naïve Bayes are predicting 51, 2, 1516 beats, respectively. Whenever we combine these three models using the voting scheme as shown in Table 12, this combined model identified 779 beats correctly. The voting scheme uses the product of probabilities rule. In this scheme, it is assumed that each model representation for a given class is statistically independent. It is because of the different representation capabilities of each model. From this, a final decision rule is formed as described in Section 2.4. This decision rule quantifies the probability of class choice from combined hypothesis models and the same type of results we can observe for other classes. In this work, we implemented four voting schemes with different classifier combinations. Each combination again provides different but better results than individual classifier models. The proposed combinations of classifier details are given in Table 27.

4.1. Comparative Analysis

To assess the performance of our proposed methodology, we compared our results with the existing methods in the literature. Comparisons are presented in Table 28 and Table 29. The features and classification schemes employed by various researchers listed in Table 28 and Table 29 are given in Table 27. In Table 28, we compare our four sets of voting schemes with the works which followed AAMI recommendations based on [31] division scheme. In addition, Table 29 shows the performance measure comparison with the literature on only N, V, and S classification. In Table 28, our proposed methods yield almost similar performance compared to the state-of-the-art for N, S, and V classes; however, in case of F and Q classes, our proposed work one and four outperforms the other compared methods. From Table 29, it is evident that our proposed methods one and three are efficiently distinguished the classes N, S, and V. The best results of our method are highlighted in bold. Overall, the measures of our work are appreciable compared with other approaches.

4.2. Limitation and Future Scope

Despite the proposed method giving significant results, the performance of the S class is still limited when compared to other classes. Similar behavior is observed in other state-of-the-art methods. Hence, there is a need to explore a new set of attributes and learning algorithms to improve this. In addition, incorporating other physiological signals such as blood pressure, plethysmographic signals along with ECG may improve the description of “heart functioning.”

5. Conclusions

In this work, we implemented a computer-aided inter-patient heartbeat classification algorithm. We employed a nonlinear decomposition method called ICEEMED, to extract some important information lying in ECG. Later, HOS and entropy measures are calculated on the modes obtained after ICEEMD and used as features. Class imbalance is one of the critical challenges in medical diagnosis. We addressed this issue by utilizing the voting scheme as the learning model. The extracted features are then fed to this model for classification. To design this model, naïve Bayes, linear and quadratic discriminating functions, J48 and J48 consolidated classifiers are explored. The proposed method showed promising results compared to state-of-the-art techniques. Our method opens new frontiers to the successful identification of rare heartbeat groups enabling a real-time heart monitoring system.

Author Contributions

Conceptualization, R.N.V.P.S.K.; methodology, K.N.V.P.S.R.; software, R.N.V.P.S.K.; validation, K.N.V.P.S.R.; formal analysis, R.N.V.P.S.K.; investigation, R.N.V.P.S.K.; resources, R.N.V.P.S.K.; data curation, R.N.V.P.S.K.; writing—original draft preparation, R.N.V.P.S.K. and R.D.; writing—review and editing, R.N.V.P.S.K., R.D., P.P., G.R.N., H.M., G.D.G. and S.G.; visualization, R.N.V.P.S.K.; supervision, R.N.V.P.S.K., R.D. and P.P.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alwan, A. Global Status Report on Noncommunicable Diseases 2010; World Health Organization: Geneva, Switzerland, 2011. [Google Scholar]
  2. Augustyniak, P.; Tadeusiewicz, R. Assessment of electrocardiogram visual interpretation strategy based on scanpath analysis. Physiol. Meas. 2006, 27, 597. [Google Scholar] [CrossRef] [PubMed]
  3. Moody, G.B.; Mark, R.G. The impact of the MIT-BIH arrhythmia database. IEEE Eng. Med. Biol. Mag. 2001, 20, 45–50. [Google Scholar] [CrossRef] [PubMed]
  4. Pławiak, P. Novel methodology of cardiac health recognition based on ECG signals and evolutionary-neural system. Expert Syst. Appl. 2018, 92, 334–349. [Google Scholar] [CrossRef]
  5. Yang, J.; Bai, Y.; Lin, F.; Liu, M.; Hou, Z.; Liu, X. A novel electrocardiogram arrhythmia classification method based on stacked sparse auto-encoders and softmax regression. Int. J. Mach. Learn. Cybern. 2018, 9, 1733–1740. [Google Scholar] [CrossRef]
  6. Tuncer, T.; Dogan, S.; Pławiak, P.; Acharya, U.R. Automated arrhythmia detection using novel hexadecimal local pattern and multilevel wavelet transform with ECG signals. Knowl. Based Syst. 2019, 104923. [Google Scholar] [CrossRef]
  7. Rajesh, K.N.; Dhuli, R. Classification of ECG heartbeats using nonlinear decomposition methods and support vector machine. Comput. Biol. Med. 2017, 87, 271–284. [Google Scholar] [CrossRef]
  8. Pławiak, P. Novel genetic ensembles of classifiers applied to myocardium dysfunction recognition based on ECG signals. Swarm Evol. Comput. 2018, 39, 192–208. [Google Scholar] [CrossRef]
  9. Yıldırım, Ö.; Pławiak, P.; Tan, R.S.; Acharya, U.R. Arrhythmia detection using deep convolutional neural network with long duration ECG signals. Comput. Biol. Med. 2018, 102, 411–420. [Google Scholar] [CrossRef]
  10. Pławiak, P.; Acharya, U.R. Novel deep genetic ensemble of classifiers for arrhythmia detection using ECG signals. Neural Comput. Appl. 2019, 1–25. [Google Scholar]
  11. Pławiak, P.; Abdar, M. Novel Methodology for Cardiac Arrhythmias Classification Based on Long-Duration ECG Signal Fragments Analysis. In Biomedical Signal Processing; Springer: Singapore, 2020; pp. 225–272. [Google Scholar]
  12. Khalaf, A.F.; Owis, M.I.; Yassine, I.A. A novel technique for cardiac arrhythmia classification using spectral correlation and support vector machines. Expert Syst. Appl. 2015, 42, 8361–8368. [Google Scholar] [CrossRef]
  13. Mert, A. ECG feature extraction based on the bandwidth properties of variational mode decomposition. Physiol. Meas. 2016, 37, 530. [Google Scholar] [CrossRef] [PubMed]
  14. Li, H.; Liang, H.; Miao, C.; Cao, L.; Feng, X.; Tang, C.; Li, E. Novel ECG signal classification based on KICA nonlinear feature extraction. Circ. Syst. Signal Process. 2016, 35, 1187–1197. [Google Scholar] [CrossRef]
  15. Alickovic, E.; Subasi, A. Medical decision support system for diagnosis of heart arrhythmia using DWT and random forests classifier. J. Med. Syst. 2016, 40, 108. [Google Scholar] [CrossRef] [PubMed]
  16. Martis, R.J.; Acharya, U.R.; Lim, C.M.; Mandana, K.; Ray, A.K.; Chakraborty, C. Application of higher order cumulant features for cardiac health diagnosis using ECG signals. Int. J. Neural Syst. 2013, 23, 1350014. [Google Scholar] [CrossRef] [PubMed]
  17. Sharma, P.; Ray, K.C. Efficient methodology for electrocardiogram beat classification. IET Signal Process. 2016, 10, 825–832. [Google Scholar] [CrossRef]
  18. Mishra, A.K.; Raghav, S. Local fractal dimension based ECG arrhythmia classification. Biomed. Signal Process. Control 2010, 5, 114–123. [Google Scholar] [CrossRef]
  19. Osowski, S.; Hoai, L.T.; Markiewicz, T. Support vector machine-based expert system for reliable heartbeat recognition. IEEE Trans. Biomed. Eng. 2004, 51, 582–589. [Google Scholar] [CrossRef]
  20. Ye, C.; Kumar, B.V.; Coimbra, M.T. Heartbeat classification using morphological and dynamic features of ECG signals. IEEE Trans. Biomed. Eng. 2012, 59, 2930–2941. [Google Scholar]
  21. Lin, C.H. Classification enhancible grey relational analysis for cardiac arrhythmias discrimination. Med. Biol. Eng. Comput. 2006, 44, 311–320. [Google Scholar] [CrossRef]
  22. Cuesta-Frau, D.; Biagetti, M.O.; Quinteiro, R.A.; Mico-Tormos, P.; Aboy, M. Unsupervised classification of ventricular extrasystoles using bounded clustering algorithms and morphology matching. Med. Biol. Eng. Comput. 2007, 45, 229–239. [Google Scholar] [CrossRef]
  23. Tadeusiewicz, R. Neural networks as a tool for modeling of biological systems. Bio-Algorithms Med.-Syst. 2015, 11, 135–144. [Google Scholar] [CrossRef]
  24. Kutlu, Y.; Kuntalp, D. A multi-stage automatic arrhythmia recognition and classification system. Comput. Biol. Med. 2011, 41, 37–45. [Google Scholar] [CrossRef] [PubMed]
  25. Martis, R.J.; Acharya, U.R.; Min, L.C. ECG beat classification using PCA, LDA, ICA and discrete wavelet transform. Biomed. Signal Process. Control 2013, 8, 437–448. [Google Scholar] [CrossRef]
  26. Martis, R.J.; Acharya, U.R.; Lim, C.M.; Suri, J.S. Characterization of ECG beats from cardiac arrhythmia using discrete cosine transform in PCA framework. Knowl.-Based Syst. 2013, 45, 76–82. [Google Scholar] [CrossRef]
  27. Elhaj, F.A.; Salim, N.; Harris, A.R.; Swee, T.T.; Ahmed, T. Arrhythmia recognition and classification using combined linear and nonlinear features of ECG signals. Comput. Methods Prog. Biomed. 2016, 127, 52–63. [Google Scholar] [CrossRef]
  28. Li, P.; Wang, Y.; He, J.; Wang, L.; Tian, Y.; Zhou, T.s.; Li, T.; Li, J.s. High-Performance Personalized Heartbeat Classification Model for Long-Term ECG Signal. IEEE Trans. Biomed. Eng. 2017, 64, 78–86. [Google Scholar]
  29. Desai, U.; Martis, R.J.; Nayak, C.G.; Sarika, K.; Seshikala, G. Machine intelligent diagnosis of ECG for arrhythmia classification using DWT, ICA and SVM techniques. In Proceedings of the India Conference (INDICON), New Delhi, India, 17–20 Deceember 2015; pp. 1–4. [Google Scholar]
  30. Desai, U.; Martis, R.J.; GURUDAS NAYAK, C.; Seshikala, G.; Sarika, K.; SHETTY K, R. Decision support system for arrhythmia beats using ECG signals with DCT, DWT and EMD methods: A comparative study. J. Mech. Med. Biol. 2016, 16, 1640012. [Google Scholar] [CrossRef]
  31. De Chazal, P.; O’Dwyer, M.; Reilly, R.B. Automatic classification of heartbeats using ECG morphology and heartbeat interval features. IEEE Trans. Biomed. Eng. 2004, 51, 1196–1206. [Google Scholar] [CrossRef]
  32. Egila, M.G.; El-Moursy, M.A.; El-Hennawy, A.E.; El-Simary, H.A.; Zaki, A. FPGA-based electrocardiography (ECG) signal analysis system using least-square linear phase finite impulse response (FIR) filter. J. Elec. Syst. Inf. Technol. 2016, 3, 513–526. [Google Scholar] [CrossRef]
  33. Raj, S.; Maurya, K.; Ray, K.C. A knowledge-based real time embedded platform for arrhythmia beat classification. Biomed. Eng. Lett. 2015, 5, 271–280. [Google Scholar] [CrossRef]
  34. Zairi, H.; Talha, M.K.; Meddah, K.; Slimane, S.O. FPGA-based system for artificial neural network arrhythmia classification. Neural Comput. Appl. 2019, 1–16. [Google Scholar] [CrossRef]
  35. Jewajinda, Y.; Chongstitvatana, P. FPGA-based online-learning using parallel genetic algorithm and neural network for ECG signal classification. In Proceedings of the ECTI-CON2010: The 2010 ECTI International Confernce on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, Chiang Mai, Thailand, 19–21 May 2010; pp. 1050–1054. [Google Scholar]
  36. Rajesh, K.N.; Dhuli, R. Classification of imbalanced ECG beats using re-sampling techniques and AdaBoost ensemble classifier. Biomed. Signal Process. Control 2018, 41, 242–254. [Google Scholar] [CrossRef]
  37. Amann, A.; Tratnig, R.; Unterkofler, K. Reliability of old and new ventricular fibrillation detection algorithms for automated external defibrillators. Biomed. Eng. Online 2005, 4, 60. [Google Scholar] [CrossRef] [PubMed]
  38. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  39. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  40. Torres, M.E.; Colominas, M.A.; Schlotthauer, G.; Flandrin, P. A complete ensemble empirical mode decomposition with adaptive noise. In Proceedings of the Acoustics, speech and signal processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 4144–4147. [Google Scholar]
  41. Colominas, M.A.; Schlotthauer, G.; Torres, M.E. Improved complete ensemble EMD: A suitable tool for biomedical signal processing. Biomed. Signal Process. Control 2014, 14, 19–29. [Google Scholar] [CrossRef]
  42. Li, T.; Zhou, M. ECG classification using wavelet packet entropy and random forests. Entropy 2016, 18, 285. [Google Scholar] [CrossRef]
  43. Rosso, O.A.; Blanco, S.; Yordanova, J.; Kolev, V.; Figliola, A.; Schürmann, M.; Başar, E. Wavelet entropy: A new tool for analysis of short duration brain electrical signals. J. Neurosci. Method. 2001, 105, 65–75. [Google Scholar] [CrossRef]
  44. Shannon, C.E. A mathematical theory of communication, Part I, Part II. Bell Syst. Tech. J. 1948, 27, 623–656. [Google Scholar] [CrossRef]
  45. Coifman, R.R.; Wickerhauser, M.V. Entropy-based algorithms for best basis selection. IEEE Trans. Inf. Theory 1992, 38, 713–718. [Google Scholar] [CrossRef]
  46. Martis, R.J.; Acharya, U.R.; Ray, A.K.; Chakraborty, C. Application of higher order cumulants to ECG signals for the cardiac health diagnosis. In Proceedings of the 2011 Annual International Conference on Engineering in Medicine and Biology Society (EMBC), Boston, MA, USA, 30 August–3 September 2011; pp. 1697–1700. [Google Scholar]
  47. Nikias, C.L.; Mendel, J.M. Signal processing with higher-order spectra. IEEE Signal Process. Mag. 1993, 10, 10–37. [Google Scholar] [CrossRef]
  48. Swami, A.; Mendel, J.M.; Nikias, C.L.M. Higher-order spectral analysis toolbox. Tech. Support Product Enhanc. Suggest. 1984. [Google Scholar]
  49. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evolut. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef] [Green Version]
  50. Kittler, J.; Hatef, M.; Duin, R.P.; Matas, J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 226–239. [Google Scholar] [CrossRef] [Green Version]
  51. John, G.H.; Langley, P. Estimating continuous distributions in Bayesian classifiers. In Proceedings of the Eleventh conference on Uncertainty in Artificial Intelligence, Montréal, QC, Canada, 18–20 August 1995; pp. 338–345. [Google Scholar]
  52. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  53. Quinlan, J.R. C4. 5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  54. Arbelaiz Gallego, O.; Gurrutxaga, I.; Lozano, F.; Muguerza, J.; Pérez, J.M. J48Consolidated: An Implementation of CTC Algorithm for WEKA. 2016. Available online: https://addi.ehu.es/handle/10810/17314 (accessed on 17 November 2019).
  55. Yang, Y.; Webb, G.I. Discretization for naive-Bayes learning: managing discretization bias and variance. Mach. Learn. 2009, 74, 39–74. [Google Scholar] [CrossRef]
  56. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2016. [Google Scholar]
  57. Powers, D.M. Evaluation: From Precision, Recall and F-Measure to ROC, Informedness, Markedness and Correlation. Available online: https://bioinfopublication.org/files/articles/2_1_1_JMLT.pdf (accessed on 19 November 2019).
  58. Luz, E.J.D.S.; Nunes, T.M.; De Albuquerque, V.H.C.; Papa, J.P.; Menotti, D. ECG arrhythmia classification based on optimum-path forest. Expert Syst. Appl. 2013, 40, 3561–3573. [Google Scholar] [CrossRef] [Green Version]
  59. Queiroz, V.; Luz, E.; Moreira, G.; Guarda, Á.; Menotti, D. Automatic cardiac arrhythmia detection and classification using vectorcardiograms and complex networks. In Proceedings of the 2015 37th Annual International Conference on Engineering in Medicine and Biology Society (EMBC), Ilan, Italy, 25–29 August 2015; pp. 5203–5206. [Google Scholar]
  60. Luz, E.J.D.S.; Merschmann, L.H.; Menotti, D.; Moreira, G.J. Evaluating a hierarchical approach for heartbeat classification from ECG. Int. J. Bioinf. Res. Appl. 2017, 13, 146–160. [Google Scholar] [CrossRef]
  61. Garcia, G.; Moreira, G.; Luz, E.; Menotti, D. Improving automatic cardiac arrhythmia classification: Joining temporal-VCG, complex networks and SVM classifier. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 3896–3900. [Google Scholar]
  62. Chen, S.; Hua, W.; Li, Z.; Li, J.; Gao, X. Heartbeat classification using projected and dynamic features of ECG signal. Biomed. Signal Process. Control 2017, 31, 165–173. [Google Scholar] [CrossRef]
  63. Garcia, G.; Moreira, G.; Menotti, D.; Luz, E. Inter-Patient ECG Heartbeat Classification with Temporal VCG Optimized by PSO. Sci. Rep. 2017, 7, 10543. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed methodology.
Figure 1. Block diagram of the proposed methodology.
Sensors 19 05079 g001
Figure 2. Scatter plot with marginal histogram for CUM2 (IMF1) vs. norm entropy (IMF1).
Figure 2. Scatter plot with marginal histogram for CUM2 (IMF1) vs. norm entropy (IMF1).
Sensors 19 05079 g002
Figure 3. Scatter plot with marginal histogram for log entropy (IMF1) vs. CUM2 (IMF1).
Figure 3. Scatter plot with marginal histogram for log entropy (IMF1) vs. CUM2 (IMF1).
Sensors 19 05079 g003
Figure 4. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF2).
Figure 4. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF2).
Sensors 19 05079 g004
Figure 5. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF3).
Figure 5. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF3).
Sensors 19 05079 g005
Figure 6. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF4).
Figure 6. Scatter plot with marginal histogram for log entropy (IMF1) vs. log entropy (IMF4).
Sensors 19 05079 g006
Figure 7. Scatter plot with marginal histogram for norm entropy (IMF1) vs. CUM2 (IMF1).
Figure 7. Scatter plot with marginal histogram for norm entropy (IMF1) vs. CUM2 (IMF1).
Sensors 19 05079 g007
Figure 8. Scatter plot with marginal histogram for norm entropy (IMF1) vs. log entropy (IMF1).
Figure 8. Scatter plot with marginal histogram for norm entropy (IMF1) vs. log entropy (IMF1).
Sensors 19 05079 g008
Table 1. Details of the number of beats selected from each group of AAMI classes.
Table 1. Details of the number of beats selected from each group of AAMI classes.
AAMI ClassesMIT-BIH HeartbeatsTotal DataTraining (DS1)Testing (DS2)
NNormal, left and right bundle branch block83,76141,74642,015
atrial and nodal escape beats
Satrial premature contraction, Aberrated atrial,26147771837
supra ventricular and junctional premature beats
Vpremature ventricular contraction, ventricular flutter689337873106
and escape beats
Ffusion of ventricular and normal526266260
beats
Qpaced, unclassifiable,1266
fusion of paced and normal beats
Table 2. Naïve Bayes Classifier parameters used in this work.
Table 2. Naïve Bayes Classifier parameters used in this work.
ParametersNaïve Bayes
Use Kernel EstimatorFalse
Use supervise DiscretizationTrue
Table 3. LDA and QDA classifier parameters used in this work.
Table 3. LDA and QDA classifier parameters used in this work.
ParametersLDAQDA
Ridge 1.0 × 10 6 1.0 × 10 6
Table 4. J48 and J48-C classifier parameters used in this work.
Table 4. J48 and J48-C classifier parameters used in this work.
ParametersJ48J48-C
Minimum Objects10001000
Use MDL correctionTrueTrue
Number of folds33
Sub-tree raisingTrueTrue
Details of the parameters can be find in WEKA 3.9 version [56].
Table 5. Confusion matrix.
Table 5. Confusion matrix.
Predicted Labels
Actua LabelsNVSFQSum
N N n N v N s N f N q R N
V V n V v V s V f V q R V
S S n S v S s S f S q R S
F F n F v F s F f F q R F
Q Q n Q v Q s Q f Q q R Q
Sum C N C V C S C F C Q R / C
Table 6. Confusion matrices for LDA, QDA, naïve Bayes, J48, and J48-C classifiers.
Table 6. Confusion matrices for LDA, QDA, naïve Bayes, J48, and J48-C classifiers.
LDAQDA
NVSFQ NVSFQ
N40,842334710930 277726636,39425780
V3192787000 53101000
S18311230 95561675110
F17100890 916420850
Q00006 06000
naïve BayesJ48
NVSFQ NVSFQ
N36,2227459973289762 41,801214000
V93018745223119 3632743000
S1321313327353 181918000
F8022455 2591000
Q10041 51000
J48-C
NVSFQ
N0694128,9665290818
V0220589461
S0241785199
F0302319116
Q00006
Table 7. Performance measures for LDA, QDA, naïve Bayes, J48 and J48-C classifiers.
Table 7. Performance measures for LDA, QDA, naïve Bayes, J48 and J48-C classifiers.
LDAQDA
OA = 92.60SEN%FPR%PPV%OA = 16.20SEN%FPR%PPV%
N97.244.694.6 6.63.793.6
V89.70.198.8 99.80.988.8
S0.10.14.1 91.280.24.4
F34.22.37.5 32.75.53.2
Q1000100 000
naïve BayesJ48
OA = 81.47SEN%FPR%PPV%OA = 94.32SEN%FPR%PPV%
N86.243.494.1 99.54794.5
V60.31.771.5 88.30.592.1
S7.22.311.2 000
F94.27.66.5 000
Q16.72.40.1 000
J48-C
OA = 8.86SEN%FPR%PPV%
N000
V71.015.924.0
S97.265.85.6
F73.511.33.5
Q1001.80.7
Table 8. Confusion matrix for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme.
Table 8. Confusion matrix for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme.
Voting (J48, LDA, naïve Bayes)
NVSFQ
N39,5425348919310
V3952708300
S14731353100
F22512320
Q00006
Table 9. Performance measures for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme.
Table 9. Performance measures for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme.
Voting (J48, LDA, naïve Bayes)
OA = 90.71SEN%FPR%PPV%
N94.136.395.4
V87.20.197.9
S19.21.141.7
F89.24.110.7
Q1000100
Table 10. Confusion matrix for combining J48, naïve Bayes, and QDA classifiers using a voting scheme.
Table 10. Confusion matrix for combining J48, naïve Bayes, and QDA classifiers using a voting scheme.
Voting ( J48, naïve Bayes, QDA)
NVSFQ
N35,629253383622970
V113095000
S118811624140
F2861111600
Q06000
Table 11. Performance measures for combining J48, naïve Bayes, and QDA classifiers using a Voting scheme.
Table 11. Performance measures for combining J48, naïve Bayes, and QDA classifiers using a Voting scheme.
Voting ( J48, naïve Bayes, QDA)
OA = 83.6SEN%FPR%PPV%
N84.823.696.7
V99.60.890.3
S348.514
F61.54.96.5
Q000
Table 12. Confusion matrix for combining J48-C, naïve Bayes, and QDA classifiers using a Voting scheme.
Table 12. Confusion matrix for combining J48-C, naïve Bayes, and QDA classifiers using a Voting scheme.
Voting (J48-C, naïve Bayes, QDA)
NVSFQ
N29,730255908924910
V83098000
S10318779190
F2161121660
Q06000
Table 13. Performance measures for combining J48-C, naïve Bayes, and QDA classifiers using a Voting scheme.
Table 13. Performance measures for combining J48-C, naïve Bayes, and QDA classifiers using a Voting scheme.
Voting (J48-C,naïve Bayes, QDA)
OA = 71.51SEN%FPR%PPV%
N70.820.396.6
V99.70.790.4
S42.420.17.9
F63.86.35.3
Q000
Table 14. Confusion matrix for combining J48-C, naïve Bayes, and LDA classifiers using a Voting scheme.
Table 14. Confusion matrix for combining J48-C, naïve Bayes, and LDA classifiers using a Voting scheme.
Voting (J48-C, naïve Bayes, LDA)
NVSFQ
N38,3296694226780
V55325322010
S14052416140
F20412350
Q00006
Table 15. Performance measures for combining J48-C, naïve Bayes, and LDA classifiers using a Voting scheme.
Table 15. Performance measures for combining J48-C, naïve Bayes, and LDA classifiers using a Voting scheme.
Voting (J48-C,naïve Bayes, LDA)
OA = 87.91SEN%FPR%PPV%
N91.23895.1
V81.50.297.2
S22.62.130.2
F90.45.78
Q1000100
Table 16. Confusion matrices for various classifiers (N, V, and S classes).
Table 16. Confusion matrices for various classifiers (N, V, and S classes).
LDAQDA
NVS NVS
N41,87527113 456948036,966
V28028260 531010
S183502 105561676
naïve BayesJ48
NVS NVS
N37,38529681662 41,998170
V922212064 109520110
S133618483 183700
J48-C
NVS
N34,99970160
V65124550
S15852520
Table 17. Performance measures for various classifiers (N,V,S classes).
Table 17. Performance measures for various classifiers (N,V,S classes).
LDAQDA
OA = 95.19SEN%FPR%PPV%OA = 19.90SEN%FPR%PPV%
N99.742.895.2 10.92.297.6
V91.00.199.1 99.81.285.3
S0.10.31.7 91.281.94.3
naïve BayesJ48
OA = 85.15SEN%FPR%PPV%OA = 93.71SEN%FPR%PPV%
N8945.794.3 10059.393.5
V68.36.841.5 64.7099.2
S26.33.821.9 000
J48-C
OA = 79.76SEN%FPR%PPV%
N83.345.294
V7916.625.2
S000
Table 18. Confusion matrix for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme (N,S,V).
Table 18. Confusion matrix for combining J48, LDA, and naïve Bayes classifiers using a Voting scheme (N,S,V).
Voting ( J48, LDA, naïve Bayes)
NVS
N40,918361736
V20528974
S14693365
Table 19. Performance measures for combining J48, LDA, and naïve Bayes classifiers using Voting scheme (N,S,V).
Table 19. Performance measures for combining J48, LDA, and naïve Bayes classifiers using Voting scheme (N,S,V).
Voting ( J48, LDA, naïve Bayes)
OA = 94.08SEN%FPR%PPV%
N97.433.996.1
V93.30.888.8
S19.91.633
Table 20. Confusion matrix for combining J48, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Table 20. Confusion matrix for combining J48, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Voting ( J48, naïve Bayes, QDA)
NVS
N37,4215744020
V1230940
S12038626
Table 21. Performance measures for combining J48, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Table 21. Performance measures for combining J48, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Voting ( J48, naïve Bayes, QDA)
OA = 87.61SEN%FPR%PPV%
N89.124.696.9
V99.61.384.2
S34.18.913.5
Table 22. Confusion matrix for combining J48-C, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Table 22. Confusion matrix for combining J48-C, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Voting ( J48-C, naïve Bayes, QDA)
NVS
N320115989406
V730990
S10628767
Table 23. Performance measures for combining J48-C, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Table 23. Performance measures for combining J48-C, naïve Bayes, and QDA classifiers using Voting scheme (N,S,V).
Voting ( J48-C, naïve Bayes, QDA)
OA = 76.40SEN%FPR%PPV%
N76.221.696.8
V99.81.483.6
S41.820.87.5
Table 24. Confusion matrix for combining J48-C, naïve Bayes, and LDA classifiers using Voting scheme (N,S,V).
Table 24. Confusion matrix for combining J48-C, naïve Bayes, and LDA classifiers using Voting scheme (N,S,V).
Voting ( J48-C, naïve Bayes, LDA)
NVS
N40,2486881079
V359273512
S14072428
Table 25. Performance measures for combining J48-C, naïve Bayes, LDA classifiers using Voting scheme (N,S,V).
Table 25. Performance measures for combining J48-C, naïve Bayes, LDA classifiers using Voting scheme (N,S,V).
Voting ( J48-C, naïve Bayes, LDA)
OA = 92.44SEN%FPR%PPV%
N95.835.795.8
V88.11.679.9
S23.32.428.2
Table 26. Median ± interquartile range of features extracted on DS1.
Table 26. Median ± interquartile range of features extracted on DS1.
Feature NumberFeaturesNSVFQ
1CUM2(IMF1)1.0 × 10 6 ± 4.0 × 10 6 8.0 × 10 6 ± 4.03 × 10 5 7.50 × 10 5 ± 0.000457750.000162 ± 0.0003260.0001405 ± 0.009121
2CUM3(IMF1)0 ± 00 ± 00 ± 2.0 × 10 6 0 ± 5.0 × 10 6 0 ± 0.000313
3CUM4(IMF1)0 ± 00 ± 01.0 × 10 6 ± 2.58 × 10 5 2.50 × 10 6 ± 1.1 × 10 5 4.5 × 10 6 ± 0.006113
4Shan(IMF1)0.004769 ± 0.0096860.017643 ± 0.065306250.112159 ± 0.41992650.216149 ± 0.3300060.153638 ± 1.47632
5log(IMF1)−4792.4962 ± 379.828−4735.1172 ± 286.978−4532.44 ± 467.507−4358.371 ± 271.824−4604.445 ± 1783.20
6norm(IMF1)0.1004 ± 0.07260.1538 ± 0.17190.0815 ± 0.057940.5541 ± 0.543418.1790 ± 20.55612
7CUM2(IMF2)0.001 ± 0.003470.00109 ± 0.00350.00028 ± 0.001370.00107 ± 0.001430.00411 ± 0.0134
8CUM3(IMF2)8.0 × 10 6 ± 5.20 × 10 5 4.0 × 10 6 ± 5.10 × 10 5 0 ± 6.0 × 10 6 5.0 × 10 6 ± 2.10 × 10 5 0 ± 0.000387
9CUM4(IMF2)1.50 × 10 5 ± 0.0001852.10 × 10 5 ± 0.0001282.0 × 10 6 ± 4.20 × 10 5 1.60 × 10 5 ± 5.20 × 10 5 0.00046 ± 0.002581
10Shan(IMF2)1.3074 ± 2.91801.3675 ± 3.42500.431 ± 1.4291.4044 ± 1.442852.90440 ± 6.76104
11log(IMF2)−4055.72 ± 403.632−4045.57 ± 565.833−4064.90 ± 479.11−3788.21 ± 266.57−4023.39 ± 1366.58
12norm(IMF2)2.41 ± 3.812.57 ± 4.952.744 ± 1.7682.69 ± 2.0418.17 ± 20.556
13CUM2(IMF3)0.0069 ± 0.00970.00534 ± 0.0090130.00443 ± 0.009360.011997 ± 0.009090.004148 ± 0.008015
14CUM3(IMF3)−7.80 × 10 5 ± 0.000322−8.60 × 10 5 ± 0.0003605−3.10 × 10 5 ± 0.000246−0.00019 ± 0.000628.95 × 10 5 ± 0.000266
15CUM4(IMF3)0.000235 ± 0.0008630.000104 ± 0.00065150.000117 ± 0.0007470.0005255 ± 0.0007670.0001945 ± 0.000396
16Shan(IMF3)6.3075515 ± 6.0758665.532014 ± 6.2548864.570122 ± 5.9390472510.002892 ± 5.492553.8205085 ± 6.888019
17log(IMF3)−3114.4106 ± 443.70730−3071.71175 ± 705.504−3178.659 ± 531.3571−2731.2660 ± 467.4692−3341.5431 ± 891.0573
18norm(IMF3)9.3564 ± 7.64448.69133 ± 8.375467.585 ± 4.146514.85 ± 7.44018.179 ± 20.55
19CUM2(IMF4)0.013817 ± 0.02100.01233 ± 0.016530.0215 ± 0.036110.0370 ± 0.02970.0069 ± 0.0102
20CUM3(IMF4)−0.00012 ± 0.00081−6.80 × 10 5 ± 0.000469−0.00014 ± 0.001305−0.001542 ± 0.002070 ± 0.00039
21CUM4(IMF4)0.000147 ± 0.0007344.60 × 10 5 ± 0.00032250.000404 ± 0.001890.00027 ± 0.001240.000138 ± 0.00018
22Shan(IMF4)13.05585 ± 12.7645712.2365 ± 11.29516.104 ± 15.13525.934 ± 15.1107.113 ± 10.471
23log(IMF4)−2150.494 ± 533.542−2143.567 ± 611.308−2206.943 ± 794.75−1614.17 ± 480.99−2593.22 ± 1271.81
24norm(IMF4)19.367 ± 15.50918.686 ± 14.24016.26 ± 9.16635.59 ± 20.94118.179 ± 20.556
25CUM2(IMF5)0.0084 ± 0.0180.0066 ± 0.01560.033 ± 0.06480.056 ± 0.0580.0075 ± 0.0149
26CUM3(IMF5)1.0 × 10 6 ± 0.000170 ± 0.000103−1.0 × 10 6 ± 0.001461.40 × 10 5 ± 0.00220 ± 0.000405
27CUM4(IMF5)−7.00 × 10 6 ± 0.00012−6.0 × 10 6 ± 0.000127−3.0 × 10 5 ± 0.00193−0.00236 ± 0.00621−5.50 × 10 6 ± 0.000287
28Shan(IMF5)10.358 ± 15.63748.68112 ± 14.82325.882 ± 29.19539.2240 ± 25.7378.992 ± 16.965
29log(IMF5)−1917.009 ± 593.83−1992.177 ± 872.920−1567.422 ± 918.993−1216.140± 489.3834−2179.500 ± 1495.143
30norm(IMF5)17.118 ± 19.073514.983 ± 19.272619.495 ± 14.65952.96 ± 32.19018.179 ± 20.556
31CUM2(IMF6)0.0040 ± 0.00920.0028 ± 0.007230.0184 ± 0.046060.0571 ± 0.08270.0073 ± 0.01072
32CUM3(IMF6)1.0 × 10 6 ± 0.000100 ± 6.80 × 10 5 2.0 × 10 6 ± 0.00111178.0 × 10 6 ± 0.005270 ± 0.0002
33CUM4(IMF6)−1.50 × 10 5 ± 0.00011-8.0 × 10 5 6± 6.25 × 10 5 −0.00024 ± 0.00205−0.00389± 0.01428−4.40 × 10 5 ± 0.00017
34Shan(IMF6)6.50405 ± 11.116975.26119 ± 9.905520.0112 ± 31.2634744.49 ± 39.513110.284 ± 13.481
35log(IMF6)−1903.169 ± 637.510−1983.32 ± 713.457−1470.196 ± 744.176−1062.5 ± 546.13−1682.363 ± 1296.591
36norm(IMF6)13.0914 ± 15.149511.4410 ± 14.26810.566 ± 12.26257.357 ± 46.31718.179 ± 20.556
Note: CUM2-second order cumulnat, CUM3-third order cumulant, CUM4-Fourth order cumulant, Shan-Shannon entropy, log-Log entropy, norm- norm entropy.
Table 27. Methodology description of recent state-of-the-art compared in our work.
Table 27. Methodology description of recent state-of-the-art compared in our work.
LiteratureFeature ExtractionClassification
[58]
method-1R–R intervalsOptimum Path Forest (OPF)
method-2Wavelet based featuresOPF
method-3Mean, standard deviation and average power of wavelet sub-bandOPF
method-4Auto correlation and energy ratio of wavelet bandsOPF
method-5Fast-ICAOPF
method-6(Wavelet+ICA+RR interval)OPF
[59](ECG+VCG) complex network based featuresSVM
[42]Wavelet packet decomposition based entropy featuresRandom Forest
[60]
method-1Wavelet based featuresHierarchical Classification (tree approach)
method-2Mean, standard deviation and average power of wavelet sub-bandHierarchical Classification (tree approach)
method-3Auto correlation and energy ratio of wavelet bandsHierarchical Classification (tree approach)
method-4Fast-ICAHierarchical Classification (tree approach)
method-5(Wavelet+ICA+RR interval)Hierarchical Classification (tree approach)
[61]Temporal Vectrcardiogram(TCG) based featuresSVM
[62]A combination of projected features
(features derived from the projected matrix and DCT) and RR intervalsSVM
[63]TCG feature selection by PSOSVM
proposed work
method-1Entropy and statistical features calculated on ICEEMD modesVoting ( J48, LDA, naïve Bayes)
method-2Entropy and statistical features calculated on ICEEMD modesVoting ( J48, QDA, naïve Bayes)
method-3Entropy and statistical features calculated on ICEEMD modesVoting ( J48-C, QDA, naïve Bayes)
method-4Entropy and statistical features calculated on ICEEMD modesVoting ( J48-C, LDA, naïve Bayes)
Table 28. Performance comparison with recent literature (N, S, V, F, and Q classes).
Table 28. Performance comparison with recent literature (N, S, V, F, and Q classes).
LiteratureNSVFQ
SEN/FPR/PPVSEN/FPR/PPVSEN/FPR/PPVSEN/FPR/PPVSEN/FPR/PPV
[58]
method-184.5/-/-1.0/-/-77.7/-/-38.4/-/-0/-/-
method-286.4/-/-2.3/-/-40.8/-/-0.5/-/-0/-/-
method-384.8/-/-18.3/-/-77.8/-/-7.5/-/-0/-/-
method-492.5/-/-3.0/-/-61.8/-/-16.8/-/-0/-/-
method-595.7/-/-17.7/-/-74.7/-/-3.9/-/-0/-/-
method-693.2/-/-12.1/-/-85.5/-/-18.3/-/-0/-/-
[59]89.3/25.2/96.638.6/6.7/1881.2/4.9/53.60/0/00/0/0
[42]94.67/3.92/99.7320/3.69/0.1694.20/0.71/89.7850/0.78/0.520/0/0
[60]
method-192.3/22.2/97.128.5/2.6/29.683.5/5.51/51.219.1/1.07/12.30/0/-
method-293.6/57.1/93.00.49/0.47/3.8167.9/3.99/54.20/1.63/00/0/0
method-398.2/41.2/95.14.72/0.71/20.381.7/1.25/82.02.58/0.40/4.880/0/0
method-498.6/39.8/95.39.15/0.56/38.683.2/1.21/82.70.26/0.38/0.530/0/-
method-594.7/31.2/96.137.4/6.19/18.843.9/1.48/67.40.52/0.72/0.560/0/-
proposed work
method-194.1/36.3/95.419.2/1.1/41.787.2/0.1/97.989.2/4.1/10.7100/0/100
method-284.8/23.6/96.734/8.5/1499.6/0.8/90.361.5/4.9/6.50/0/0
method-370.8/20.3/96.642.4/20.1/7.999.7/0.7/90.463.8/6.3/5.30/0/0
method-491.2/38/95.122.6/2.1/30.281.5/0.2/97.290.4/5.7/8100/0/100
Table 29. Performance comparison with recent literature (N, S, and V classes).
Table 29. Performance comparison with recent literature (N, S, and V classes).
LiteratureNSV
SEN/FPR/PPVSEN/FPR/PPVSEN/FPR/PPV
[61]95/27.9/96.529.6/3.1/26.485.1/3.01/66.3
[62]98.4/-/95.429.5/-/38.470.8/-/85.1
[63]
method on VCG79.1/27.0/96.331.2/8.4/13.089.5/7.2/46.1
proposed work
method-197.4/33.9/96.119.9/1.6/3393.3/0.8/88.8
method-289.1/24.6/96.934.1/8.9/13.599.6/1.3/84.2
method-376.2/21.6/96.841.8/20.8/7.599.8/1.4/83.6
method-495.8/35.7/95.823.3/2.4/28.288.1/1.6/79.9

Share and Cite

MDPI and ACS Style

Kandala, R.N.V.P.S.; Dhuli, R.; Pławiak, P.; Naik, G.R.; Moeinzadeh, H.; Gargiulo, G.D.; Gunnam, S. Towards Real-Time Heartbeat Classification: Evaluation of Nonlinear Morphological Features and Voting Method. Sensors 2019, 19, 5079. https://0-doi-org.brum.beds.ac.uk/10.3390/s19235079

AMA Style

Kandala RNVPS, Dhuli R, Pławiak P, Naik GR, Moeinzadeh H, Gargiulo GD, Gunnam S. Towards Real-Time Heartbeat Classification: Evaluation of Nonlinear Morphological Features and Voting Method. Sensors. 2019; 19(23):5079. https://0-doi-org.brum.beds.ac.uk/10.3390/s19235079

Chicago/Turabian Style

Kandala, Rajesh N V P S, Ravindra Dhuli, Paweł Pławiak, Ganesh R. Naik, Hossein Moeinzadeh, Gaetano D. Gargiulo, and Suryanarayana Gunnam. 2019. "Towards Real-Time Heartbeat Classification: Evaluation of Nonlinear Morphological Features and Voting Method" Sensors 19, no. 23: 5079. https://0-doi-org.brum.beds.ac.uk/10.3390/s19235079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop