Next Article in Journal
Transfer Information Assessment in Diagnosis of Vasovagal Syncope Using Transfer Entropy
Next Article in Special Issue
First-Stage Prostate Cancer Identification on Histopathological Images: Hand-Driven versus Automatic Learning
Previous Article in Journal
CDCS: Cluster-Based Distributed Compressed Sensing to Facilitate QoS Routing in Cognitive Video Sensor Networks
Previous Article in Special Issue
Combination of Global Features for the Automatic Quality Assessment of Retinal Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of the KNN Supervised Classification Algorithm as a Support Tool for the Implantation of Deep Brain Stimulators in Patients with Parkinson’s Disease

by
Gabriel Martin Bellino
1,*,
Luciano Schiaffino
1,2,*,
Marisa Battisti
1,3,
Juan Guerrero
4 and
Alfredo Rosado-Muñoz
4
1
Faculty of Engineering, National University of Entre Ríos: route 11, Km 10, 3100 Oro Verde, Argentina
2
Faculty of Life and Health Sciences, Autonomous University of Entre Ríos, Alem 217, 3100 Paraná, Argentina
3
Faculty of Science and Technology, Autonomous University of Entre Ríos, route 11, Km 10.5, 3100 Oro Verde, Argentina
4
GDDP, Department of Electronic Engineering, School of Engineering, Universitat de Valencia, Burjassot, 46100 Valencia, Spain
*
Authors to whom correspondence should be addressed.
Submission received: 28 February 2019 / Revised: 21 March 2019 / Accepted: 26 March 2019 / Published: 29 March 2019

Abstract

:
Deep Brain Stimulation (DBS) of the Subthalamic Nuclei (STN) is the most used surgical treatment to improve motor skills in patients with Parkinson’s Disease (PD) who do not adequately respond to pharmacological treatment, or have related side effects. During surgery for the implantation of a DBS system, signals are obtained through microelectrodes recordings (MER) at different depths of the brain. These signals are analyzed by neurophysiologists to detect the entry and exit of the STN region, as well as the optimal depth for electrode implantation. In the present work, a classification model is developed and supervised by the K-nearest neighbour algorithm (KNN), which is automatically trained from the 18 temporal features of MER registers of 14 patients with PD in order to provide a clinical support tool during DBS surgery. We investigate the effect of different standardizations of the generated database, the optimal definition of KNN configuration parameters, and the selection of features that maximize KNN performance. The results indicated that KNN trained with data that was standardized per cerebral hemisphere and per patient presented the best performance, achieving an accuracy of 94.35% (p < 0.001). By using feature selection algorithms, it was possible to achieve 93.5% in accuracy in selecting a subset of six features, improving computation time while processing in real time.

1. Introduction

Parkinson’s disease (PD) is a chronic neurodegenerative disorder that is produced by the progressive death of the neurons producing dopamine in the substantia nigra. A decrease in dopamine levels produces an effect in other basal nuclei of the brain, such as the striatum, the subthalamic nucleus, and the globus pallidus, which play an important role in the inhibition and control of the human movements [1,2]. PD is a major health problem that affects approximately 1–2% of people over 65 years of age worldwide [3]. The motor symptoms of PD are bradykinesia, tremor at rest, rigidity, and postural instability. Non-motors symptoms are sleep disorders, cognitive dysfunctions, among others [1,2].
The initial treatment of PD is pharmacological; however, in some cases, an adequate control of the symptoms is not achieved. In other cases, the medication causes undesired side effects, such as the appearance of dyskinesia or intolerance. Also, after five or six years of this treatment, its effectiveness begins to diminish and the initial symptoms reappear [1,2]. In these cases, deep brain stimulation (DBS) is an effective treatment [4]. DBS is a therapy that involves the release of an electrical current in a deep part of the brain to treat a neurological dysfunction. This electrical current is formed by pulses of controllable amplitude, frequency, and voltage, which are provided by an implantable pulse generator and transmitted through a stimulation electrode, which acts as a pacemaker of the brain and it is effective in improving the motor disorders of PD. There are two deep brain structures whose stimulation has shown to have consistent therapeutic effects on the motor symptoms of the disease and dyskinesias that are produced by levodopa. These target structures are the subthalamic nucleus (STN) and the globus pallidus in its internal segment (GPi). However, the most common stimulation zone is the STN of both cerebral hemispheres to reduce the chronic hyperactivity of the neurons involved [5,6]. Stimulation of the STN allows for reducing the doses of pharmacological treatment of the patients and it requires less energy than the stimulation of the GPi. The STN is a discoid structure of approximately 5.9 × 3.7 × 5 mm3 that is located between the diencephalon and mesencephalon surrounded by substantia nigra, red nucleus, and zona incerta [2].
A multidisciplinary team with experience in the field is required to perform the implantation of electrodes, which is ideally formed by neurologists, neurophysiologists, neurosurgeons, and trained surgery personnel. Stereotactic procedures are used in order to achieve correct implantation, being combined with magnetic resonance imaging (MRI) and computed tomography (CT) during the procedure if the technology is available, and microelectrode recordings (MER) [6,7]. Normally, the patient is not anaesthetized to undergo surgery, in some cases, only sedation is needed, since it is necessary to evaluate the clinical response to the stimulation and reduce the side effects. A macro-stimulation device is used that releases a current in the target structure reached by the electrodes to perform the stimulation during the surgical procedure.
The analysis of the MER records is a widely used method of localization in deep structures, through visual and acoustic analysis and beta rhythm analysis [4]. MER electrodes has less than 100 micrometers diameter and it goes into the brain through different functional structures, such as the anterior thalamus (TH), zona incerta (ZI), subthalamic nucleus (STN), and substantia nigra pars reticulata (SNr) [4]. Each one of them presents a specific neural activity, as seen in Figure 1, constituting non-stationary signals [4,6], conformed by: the neuronal activity called spikes, the background neuronal activity, and artifacts [8].
The localization of the STN for implantation of DBS electrodes is complex task for neurophysiologists, as presented in [4,9], a combination of anatomic localization via stereotactic procedures with functional localization via MER is used to achieve a correct implantation as well as to identify the best implantation zone in the STN. It has been proven that MER readings can reduce targeting errors that are related to the resolution limitations of the MRI and CT images, and the anatomic shifts during surgery [9]. In this context, a brain shift is produced when the burr holes are made [6]. The MERs signals amplitudes can vary from 50 µV up to 200 µV [8,9,10], and they are visually and acoustically analyzed, which requires a clean and crisp signal. Accordingly, signal conditioning must be done before starting the records, while using the pre-amplification module and software gain and filters in the set-up stage. In this stage, environmental interference or noise issues are solved.
In the last 10 years, several works have been reported [4,11], where signal processing and data mining have been used to locate specific regions of the brain, such as TH, ZI, STN, or SNr, which are solely based on MER information.
Chaovalitwongse et al. [4] proposed three classifiers: Naive Bayes, k-nearest neighbors (KNN), and decision trees, which all use 13 temporal features estimates from MER. In this work, the authors obtained 82.2% accuracy for the KNN algorithm, 82.8% accuracy with the Bayesian kernel, and 89.6% accuracy with the classification trees. In addition, a time-frequency indicator was used to train KNN and decision trees, but the obtained results were not encouraging (50.2% average accuracy). The Bayesian approach takes into account the posterior conditional probability that was calculated from each and every feature. Thus, noisy features can play a role in reducing classifier accuracy. In the classification trees algorithm, the variables and thresholds that were used at each branching point are independent of each other. This ensures that the algorithm always chooses the variable that is best suited for partitioning the set. The algorithm prioritizes the importance of each feature based on the hierarchical structure of decision-making. The result of [4] confirm that identifying the most informative features in distinguishing different subcortical structures is necessary, but the authors do not explore feature selection algorithms to choose which ones are more suitable for their classification process.
Rajpurohit et al. [12] worked with four classification methods: Logistic Regression (LR), Gaussian Naive Bayes (GNB), KNN, and support vector machine (SVM). In this case, the author and colleagues extracted 13 temporal features from 65 MER tracks for classifier training and then used sequential forward and backward feature selection. Of the 13 temporal features, six were spike-detection-independent and seven were spike-detection-dependent. When using the entire set of normalized features with patient-independent normalization, the baseline classification error rate was approximately 17–19% for each of the four supervised learning classifiers. The resulting accuracy from each classification method was: LR 83%, GNB 84%, KNN 81%, and SVM 83%. The Logistic Regression algorithm does not show different weights to the characteristics according to their relevance and this can limit the performance of the classifier. The SVM classifier takes a long time to train but it achieves good results with the linear kernel. Occasionally, the SVM algorithm presents non-convergence problems.
In [4,12], the KNN algorithm was used, but only the Euclidean distance is explored. Likewise, no study is carried out to determine the optimal value of the number of neighbors in order to improve the performance of the classifier.
Cagnan et al. [11] proposed a classification method with a structure that was similar to a binary decision tree, working with two temporal features and two frequency variables that were linked to the power in the beta and gamma frequency range, with the aim of detecting the entry and exit of the electrode in STN. The algorithm achieved 88% accuracy in the detection of STN according to the classifications that were performed by neurophysiologists.
The works described above [4,11,12] used a debugged database without noisy records for the validation process of the classifiers. Consequently, the reported results are likely to yield higher performance values than the results that were obtained with those algorithms while using real MER records, as directly obtained in a DBS surgery.
In exploratory tests that were carried out with individual supervised classifiers, the KNN algorithm presented the best performance with our database. In this article, we present a KNN based algorithm that was trained and validated with different features that were calculated from MER records, in order to identify the STN as a subcortical structure target for DBS in PD. We explore how to optimize the performance of KNN in order to achieve clinical relevance for the application of the algorithm as a support tool in real time during DBS surgeries. For the clinical purpose sought, our hypothesis suggests that the optimization of the KNN classifier can be achieved with the combination of three aspects: the different standardizations of the calculated features, the optimal definition of KNN configuration parameters, and the application of algorithms for adequate feature selection.

2. Materials and Methods

2.1. Database

The records of neuronal electrical activity at different depths of the brain were obtained from bilateral DBS surgery that was performed in 14 non-medicated patients with PD, aging 57 ± 6 (eight male/six female), where the stimulator was implanted in the STN. These surgeries were performed at the Hospital La Fe in Valencia, Spain. All of the patients met medical accepted selection criteria and signed an informed consent for DBS surgery with MER. These investigations that use human data were carried out following the rules of the Declaration of Helsinki. The Ethical Committee for Biomedical Research of La Fe Hospital approved the research procedures for the project with registration number 2015/0824 in May 17, 2016.
For surgical planning, T1 and T2 weighted 1.5T MRI series were fused with a CT scan that was performed after placing the stereotactic frame. As per standard clinical protocol, the target coordinates and trajectory to the STN were identified while using the fused images on a neuro-navigational platform (StealthStation, Medtronic Corp, Minneapolis, MN, USA). Two neurophysiologists continuously analyzed the electrical records that were obtained with MERs as the electrode goes through different structures of the brain, followed by standard stereotactic techniques. This procedure allowed for determining whether the brain structure is STN, defining the entry and exit limits, as well as identifying the best implantation zone within this basal structure. This classification was validated off-line by fused images of the intraoperative CT with the preoperative images. In our work, the record of neuronal electrical activity was performed for each cerebral hemisphere, with two platinum-iridium (Pt-Ir) MER per hemisphere that entered parallel and were separated by 2 mm each other. Medtronic MER with an impedance of 1 MΩ at 1000 Hz was used. The records were obtained at a sampling frequency of 12 kHz with a 16-bit converter, a total gain of 10,000, applying a line filter at 50 Hz, and a Butterworth band pass filter between 200 Hz and 6000 Hz. All of the records were obtained with the Alpha Omega MicroGuide ProTM system. This system also acquired the depth of the MER in [mm] respect to the surface of the skull and the position that was marked as target. Records started 7 mm before the target area with an advance of the MER electrodes in steps of 0.2 mm, recording at least 16 s in each position.

2.2. Features Obtained from MER Records

Signals were processed in 4 s windows that overlapped 50% in order to calculate 18 temporal features, in an off-line analysis using Matlab® version R2017a. It was demonstrated in [4] that this window size is optimal in capturing enough spikes to detect changes in the subcortical structures and small enough for real-time processing. The records were used as they were exported from the Alpha Omega MicroGuide ProTM equipment (Nazareth, Israel). Only the records that were acquired during the time when the MER was descending from one position to another were eliminated, since they were heavily noisy. For valid records, the following features were calculated as proposed in the state of art [4,13]:
Spike-detection-independent features:
  • Basal amplitude value (1-VAB): Dolan et al. [14] proposed a robust method using the Hilbert transform to estimate the envelope of a time record MER using Equation (1), where E ( t ) is the envelope and H { X ( t ) } is the Hilbert transform of the signal X ( t ) .
    E ( t ) = X ( t ) 2 + ( H T { X ( t ) } 2   ,
  • Root mean square amplitude (2-RMS): Calculated according to Equation (2) [4].
    R A = i = 1 n ( X i ) 2 n   ,
  • Kurtosis (3-KUR): Statistically, it is a measure that is used to describe a distribution. Whereas, skewness differentiates the extreme values in one versus the other tail, kurtosis measures extreme values in either tail. If the distributions of the variables are not known and discrete data is available, K can be estimated according to Equation (3), where μ 4 is the fourth moment with respect to the mean and σ the standard deviation.
    K = μ 4 σ 2   ,
  • Curve length (4-CL): Calculated according to Equation (4) [4]. Where x i is ith point data vector of length n .
    C L = i = 1 n 1 | X i + 1 X i |   ,
  • Threshold (5-TH): Calculated according to Equation (5) [4]. Where μ is the mean of the data vector.
    T H = 3 n 1 i = 1 n ( X i μ ) 2   ,
  • Peak count (6-PK): Calculated according to Equation (6) [4].
    P K = 1 2 i = 1 n 2 max { 0 , s g n [ X i + 2 X i + 1 ] s g n [ X i + 1 X i ] }   ,
  • Average nonlinear energy (7-NE): Calculated according to Equation (7) [4]. High values of NE show the presence of high frequency signals in the segment’s analysis.
    N E = 1 n 2 i = 2 n 1 [ X i 2 X i 1 X i + 1   ]   ,
  • Zero crossings (8-ZC): Calculated according to Equation (8) [4].
    Z C = 1 2 i = 1 n 1 s g n   ( X i + 1 ) s g n ( X i )     ,
Spike-detection-dependent features:
  • Spike burst index (9-SBI): ratio of the number inter-spike intervals (ISIs) less than 10 ms to the number that is greater than 10 ms [15].
  • Spike pause index (10-SPI): ratio of the number of ISIs greater than 50 ms to the number less than 50 ms [15].
  • Spike pause ratio (11-SPR): ratio of cumulative time of ISIss greater than 50 ms to the cumulative time of those less than 50 ms [15].
  • Spike count (12-SC): instead of using the spike count, in this work, we proposed to use spike frequency, in spike per seconds of the segments analyzed.
  • Mean spike amplitude differential (13-SMAD): 80 percent trimmed mean of the difference between consecutive spike amplitudes [4].
  • Spike count ratio (14-SCR): fraction percentage of spikes accepted as genuine spikes among candidate spikes by the spike detector [4].
  • Median of the spike count (15-SF): Calculated as the median of SC.
  • Standard deviation of the ISIs (16-SSD) [4].
  • Mean value of the ISIs (17-SDp).
  • Median value of the ISIs (18-SDm).
The database was formed with 34,898 records of 4 s windows, in which we calculated the 18 features listed above, and each window was labeled as STN or non-STN according to the data that was provided by trained neurophysiologists and co-register images. In this context, 52% of the total windows corresponded to the STN class. From this database, three new databases were generated with the same number of records as the original in which different standardizations were applied to each feature. Subsequently, a standardization of the original database was made, subtracting the mean and dividing by the standard deviation of the features that were selected by these criteria:
  • Each feature was standardized based on the values of that feature calculated for all patients.
  • Each feature for patient was standardized based on the values of that feature calculated for own patient.
  • Each feature for patient and hemisphere was standardized based on the values of that feature calculated for own patient and each cerebral hemisphere.

2.3. K-Nearest Neighbors Classifier (KNN)

Let R(z) ⊂ ℜN be a hypersphere with volume V and center z. Nk is the number of samples of the training set Tk for the classifier KNN and Wk the class assigned. The probability of having exactly n samples within R(z) has a binomial distribution [16], according to (9).
E [ n ] = N k y   R ( z ) p ( y   |   w k )   d y   N k V p ( y | w k )     ,
If a radius is set around z that generates a volume that contains exactly K samples, then that radius and its volume depend on the position z in the measurement space [16]. Therefore, we can write V(z) instead of V being the estimate of density [17], as indicated in (10).
p ^ ( z   |   w k ) = K N k V ( z )     ,
In the KNN algorithm, the K value is set to estimate the model and the minimum volume V(z) that these K samples cover is calculated. The expression (2) indicates that, in the regions where the density estimate is large, the volume is expected to be small. If the estimate is small, then the sphere needs to grow to collect the necessary samples [17]. Moreover, the K parameter controls the balance between bias and variance, as indicated in (11) [16].
K     y   N k     ,   in   order   to   get   a   small   variance ; K N k     0   y   N k     ,   in   order   to   get   a   small   bias
The KNN technique has a practical interest, since it works on the set of samples to estimate the model without calculating the probability density. If Kk is the number of neighboring samples that are found in the Wk class, then a conditional density estimator is (12).
p ^ ( z   |   w k ) K N k V ( z )
By combining (2.12) with the Bayes Naive classifier with a uniform cost function [16], it is possible to obtain the estimated classification according to (13).
w ^ ( z ) = w k ; k = arg max i = 1 , , k { p ^ ( z | w i ) P ^ ( w i ) } ; k = arg max i = 1 , , k { K i N i V ( z ) N i N s   } ; k = arg max i = 1 , , k { K i }
From the previous expression, it is concluded that the class that is assigned to the vector z is that with the largest number of neighbour samples of the class Wk closest to z. In this work, Bayesian optimization is applied to minimize the classification error with our database, and thus to determine the optimal configuration of KNN. In this work, a value of K = 9 and a city block distance metric were adopted for KNN algorithm.

2.4. Performance of KNN Classifiers

For the training and validation process, 14 different subsets were generated for each of four databases (original and three with different standardization, as described in Section 2.2) in order to simulate a real situation as during a DBS surgery. In each subset, subset 1, for example, the data from patient 1 was taken for validation and data from the rest of the patients (2 to 14) for training and then, for the other subsets, the same procedure was applied, as proposed in [4]. This method is known as leaving one patient out. In this way, 14 KNN classifiers were calculated with each database with non-standardized (KNN) features, and standardized features: with data from all patients (KNN_STA); with data per patient (KNN_PAT); and, with data per patient and per cerebral hemisphere (KNN_HEM). Subsequently, using each respective validation datasets, the performance of the classifiers were analyzed by calculating the performance indices that were widely reported in the state of art [4,12,18]: accuracy (ACC), sensitivity (SEN), specificity (ESP), area under the ROC curve (AUC), and index of diagnosis (DOR).
A descriptive statistic study was carried out and statistical comparisons were made with nonparametric tests, after checking for assumptions of normality (Kolmogorov–Smirnov test; p < 0.05 ). Friedman test was used for the analysis of global significance, and for the paired comparisons of classifiers, Nemenyi was used as a post-hoc test. The threshold of significance between the comparisons was accepted at 95% ( p < 0.05 ). All of the results are expressed as mean and standard deviation (Mean ± SD). SPSS Statistics v24 and Statistics and Machine Learning toolbox for Matlab® version R2017a were used to calculate the performance indexes and statistical analysis.

2.5. Feature Selection Techniques

Feature selection for classification can be defined as a combinatorial optimization problem, where a set of features is selected in a way that maximizes the quality of the hypothesis that was learned from these features. Supervised methods of feature selection can be categorized in filter models and wrapper models [19], where the methods are not intrinsic to the classification algorithm.
Filter models separate the feature selection from the classifier learning, so that the bias of a learning algorithm does not interact with the bias of a feature selection algorithm [20]. It is based on measurements of the general characteristics of training data, such as distance, consistency, dependence, information, and correlation. These methods can be univariate if they do not take the values of other attributes when the selection procedure took place into account, or multivariate if they consider the interaction of the other characteristics [20]. Robnik-Sikonj et al. [21] related ReliefF’ s relevance evaluation criterion to the margin maximization hypothesis, concluding that ReliefF provides superior performance in many applications, being a more stable algorithm than other filter type algorithms [22]. ReliefF (RS), being multivariate, also allows for detecting redundant features, which is why it was selected in the present work as a filter type selection method [22].
Kononenko et al. in 1994 proposed relief and RS class extension [21]. Basically, the method consists of selecting features randomly and then, depending on the closest neighbors, assigning more weight to the features that better discriminate among classes. When considering that instances are randomly sampled from data, then the score of the i-th feature Si is defined according to Equation (14), where Mk denotes the values in the i-th feature of the closest instances Xk with same class tag, while Hk denotes the values in the i-th function of the instances that are closest to Xk with different class tags, and d(·) is a measure of distance [21].
S i = 1 2 k = 1 l [ d ( X i k X i M k ) d ( X i k X i H k ) ]
In the wrapper model, the searching procedure for subsets of features consists in generating several subsets that were obtained from the original set that are evaluated while using a classifier algorithm [22]. It is important to note that the number of subsets increases exponentially as the number of features increases, and it is necessary to use heuristic methods to guide the search, making these methods somewhat challenging [22]. Generally, wrapping methods have better performance than filter methods [19], since improving the classification algorithm also improves the process of feature selection. However, wrapping methods can be computationally more expensive for problems with large dimensions, greater than 50 features, since each subset of features considered must be evaluated by the classification algorithm [22]. In this work, we used three wrapper model algorithms that were applied to the KNN classifiers described in the previous sections: backward, forward, and branch and bound. Once the classifier is selected, a wrapping model will perform the following steps:
  • Step 1: Look for a subset of features,
  • Step 2: Evaluate the subset of features selected by the performance of the KNN classifier, and
  • Step 3: Repeat Step 1 and Step 2 until reaching the desired performance.
The feature search methodology produces a subset of features that are used to train the classifier. The resulting classifier is then evaluated with an independent data set that has not been used in the training process [22]. As mentioned before, several heuristic search strategies were used, such as: backward elimination, forward selection, and branch and bound [19]. In the forward technique, search begins with an empty set of features and then progressively incorporates features into a larger subset. In the backward technique, it begins with a set that conformed to all features and those that do not contribute to the performance of the classifier are progressively eliminated. In branch and bound method, the algorithm starts from the full set and it removes the features through a first deep search with a reverse strategy. In this last strategy, a search is systematically carried out by means of a tree structure when considering that nodes whose objective function are lower than the current best ones are not explored, under the assumption of monotony that ensures that their sons will not contain a better solution. The branch and bound method (BBS) has shown that it does not fall into local minimums, such as backward (BS) and forward (FS), given the algorithm’s search nature [19]. The feature selection algorithms that were described in this section were applied to one database selected based on best performance indicators.

2.6. Performance of KNN Classifiers with Feature Selection

The same sets of training and testing data that are detailed in Section 2.4 were used for KNN classification with all features (KNN) and with the features resulting after applying feature selection methods, such as ReliefF (KNN+RS), backward (KNN+BS), forward (KNN+FS), and branch and bound (KNN+BBS). Thus, the results that were obtained of each classifier are comparable.
We analyzed KNN+RS, KNN+BS, KNN+FS KNN+BBS, and KNN, as five different classifiers. For each classifier, the performance indices detailed in Section 2.4 were calculated: SEN, ESP, ACC, and AUC. We carried out the non-parametric Friedman test, followed by Nemenyi post-hoc test for pairwise comparisons if the results of the Friedman test indicated overall significance ( p < 0.05 ), as presented in Section 2.4.

3. Results and Discussion

3.1. Results of the KNN Classifiers

Table 1 shows the average results that were obtained from the 14 classifiers for the four versions of the KNN algorithm according to Section 2.4. Likewise, Table 2 presents the average training and validation times. The results indicate that all four classifiers train quickly and at similar times. On the other hand, the validation process for the data of a new patient represents an order of magnitude higher than the training time.
Figure 2 presents the ROC curves for the four versions of the algorithm. It can be seen that both the average values of Table 1 and the area under the ROC curve of Figure 2 show that the KNN_HEM version presents the best average performance for all of the indicators.
The obtained results for the four trained classifiers were statistically compared by Friedman and each performance indicator. Given that, in all cases, the test showed global significance ( p < 0.001 ) and the Nemenyi test was performed to obtain pairwise comparisons, whose results are presented in Table 3.
The standardization of features by patient and by hemisphere (KNN-HEM) allowed for raising the mean accuracy by 15% when compared with the KNN that was trained with the database without standardized (KNN) and 10% higher when compared with KNN trained with the database standardized of all patients (KNN-STA). From the analysis of the previous results, it is observed that, despite all the trained KNN classifiers showing an acceptable performance, KNN_HEM presents a significant improvement when compared with KNN ( p < 0.001 ) and KNN_EST ( p < 0.001 ). This fact shows that, even under the experimental conditions as the present work, standardizing the data per patient and per cerebral hemisphere allows for us to obtain encouraging results. Analyzing the increase in performance indices according to the standardization method that is used reflects this improvement.
Our study obtained 94.35% mean accuracy while using the KNN-HEM algorithm, while Rajpurohit et al. [12] obtained 81% accuracy and Chaovalitwongse et al. [4] obtained 82.2% accuracy using KNN. In addition, the best accuracy values that were obtained in Rajpurohit et al. [12] were obtained with the GNB classifier, obtaining an average accuracy of 84%. In the work by Rajpurohit et al. [12], in which 10 cross-validation was used as a data partition for training and testing, bias is very likely to occur in favor of the performance accuracy of the classifiers being used. Chaovalitwongse et al. [4] obtained the best performance with a classifier tree, obtaining an average accuracy of 89.6%
Although it is not possible to make a direct comparison with the results that were obtained in other works, since different databases are used, the classification percentages obtained in the present work with KNN_HEM are higher than those that were obtained by [4,11]. It is observed from Table 1 that, in all cases, the sensitivity is superior to the specificity, which indicates that the KNN algorithm has greater capacity to detect the STN area, which presents clinical relevance in the context of the application.

3.2. Results of the KNN Classifiers with Feature Selection

Table 4 and Table 5 present the selected features for each feature selection algorithms, showing the results of the performance indicators for each classifier trained with selected features for each algorithm along with the KNN trained with all features. As presented, KNN+RS and KNN+BBS trained with seven and six features respectively, of which five of them are the same, accuracy, as other indicators, are slightly lower than KNN (0.98% for KNN+RS and 0.9% for KNN+BBS). It is noteworthy that the number of features is one-third of the 18 features, which facilitates the process of classification in real time. KNN+BS (trained with 12 features) and KNN+FS (trained with 11 features) slightly improve the performance indicators when compared with KNN, 1.21% and 1.13%, respectively, higher for accuracy.
Table 6 presents the results of Friedman and Nemenyi tests for each classifier with feature selection and for each performance indicator. It can be observed that, statistically, KNN+BS and KNN+RS models present significant differences for all of the indicators with KNN+RS and KNN+BBS. KNN+FS does not present statistically significant differences for any performance indicator when compared with KNN. In the case of KNN+BS, there are two indicators that show statistically significant differences with KNN: ACC ( p = 0.047 ) and SEN ( p = 0.013 ).
The performance indicators for the different models that were obtained with feature selection do not present average percentage differences that indicate a clinical relevance. From this perspective, it is more important to achieve models that require a smaller number of features to be calculated during a surgery, and thus guarantee a classification in a shorter time. In the case of the models with fewer features, (KNN + RS and KNN + BBS), KNN + BBS has a higher value than KNN + RS for three of the four performance indicators, and at the same time, KNN + BBS does not present statistically significant differences with KNN for all of the indicators. The branch and bound method managed to select this subset of features that is present in all other feature selection algorithms, resulting in an adequate combination of four features that are linked to the background activity (4-CL, 5-TH, 6-PK, and 8-ZC) and two linked to the spikes (13-SMAD and 14-SCR). These results are concordant with that obtained by Rajpurohit et al. [12], which, when applying the features selection by wrapping methods (backward and forward), managed to improve the performance of the KNN classifiers, logistic regression, SVM, and Bayesian developed in their work. In the case of KNN, which has the lower error in this work, it was trained with seven features out of a total of 13, five of them being related to the background activity spike-detection-independent and two features that are related to the activity of spikes.
Other authors, such as Novak et al. [9], reported that the background activity in the STN of 15 patients presented a statistically significant difference between the STN, TH, and ZI, as well as with substantia nigra. In this paper, they also reported that the TH amplitude of background activity is similar to the substantia nigra. Przybyszewski et al. [23] reported that, in a study with 10 patients with statistical validation, the background activity of multiple unit activity in combination with the average power in the beta range of local field potentials had discriminating power to detect the STN of nearby structures. In both studies, it was concluded that having features linked to the background activity is important in the detection of STN.
In regards to feature selection techniques, all of the selection methods used conferred more importance to the features that are associated with the background activity. In particular, 4-CL, 5-TH, 6-PK, and 8-ZC are present in all of the results obtained by the different feature selection methods, possibly those that best represent the background activity. At the same time, features 13-SMAD and 14-SCR, which are associated with the spikes, were also selected by all methods, showing that they better characterize the activity of the spikes.

4. Conclusions

The results of the present work initially propose a KNN_HEM model that constitutes a first step for an automatic classification system that works in the operating room as a support tool for neurophysiologists and neurosurgeons when defining the optimal location in the fixation of the stimulation electrode of a DBS system in patients with PD. The computational times that were obtained in training and validation imply that a KNN algorithm can be used to validate the data of a new patient in real time. A system of these characteristics will allow for the reduction of times of a surgery of this nature, providing an objective classification result.
Initially, 18 temporal features reported in the state of the art were adopted [4,11,12], with good results in the supervised classification. Later, investigating the algorithms of feature selection to those that could optimize the classification process with the data available in the database that was used is a future direction. The feature selection methods that were addressed in the present work allowed for obtaining classifiers with a performance that is similar or superior to the classifiers trained with all of the characteristics, eliminating those redundant or noisy features. It is noteworthy that obtaining trained classifiers with a smaller number of characteristics not only improves the classification time of the validation data, but also the calculation time of the characteristics from the MER records, thus improving the full computation time of the whole process.
Background activity represents small action potentials, with noise order levels, which are generated by neural populations near the recording electrode but not in direct contact with it, this describes the neuronal behavior of an STN volume. This could characterize neuronal population and allows, through supervised classification, the detection of registers coming from that basal nucleus. Neurophysiologists also identify this zone based on the spikes, analyzing both their amplitude and firing frequencies, which correspond to the action potentials of a group of neurons that is in contact with the electrode.
There is a lower capacity of the KNN algorithm as well as those that were reported by other authors [11,12] to differentiate the STN based on this information. Based on the results obtained and under the experimental conditions of the present work, it is proposed to use KNN + BBS in the classification process in real time during DBS surgeries.

Author Contributions

G.M.B. and L.S. trained and optimized the classifiers, analyzed the data, interpreted the results, and wrote the manuscript. L.S. took part in the collection of the database. M.B. applied feature selection algorithms and interpreted the results. J.G. processed the MER records and performed the feature calculation. J.G. and A.R.-M. designed the study and coordinated group activities. All authors have read, reviewed, and approved the manuscript.

Funding

This research was funded by the National University of Entre Rios (UNER—research project number 6169), by the Autonomous University of Entre Rios (UADER—Res CS 067/2018) and by University of Valencia (VLC-BIOMED 2015 Subprogram B, project 02-Parkinson-2015-B).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Peñas Domingo, E.; Sierra, M.G.; Valero, M.M.; Castiñeira, M.P.-O. El libro blanco del Parkinson en España; Federación Española de Párkinson: Madrid, Spain, 2015. [Google Scholar]
  2. Schapira, A.H.; Jenner, P. Etiology and pathogenesis of Parkinson’s disease. Mov. Disord. 2011, 26, 1049–1055. [Google Scholar] [CrossRef] [PubMed]
  3. Dorsey, E.R.; Constantinescu, R.; Thompson, J.P.; Biglan, K.M.; Holloway, R.G.; Kieburtz, K.; Marshall, F.J.; Ravina, B.M.; Schifitto, G.; Siderowf, A.; et al. Projected number of people with CME Parkinson disease in the most populous nations, 2005 through 2030. Neurology 2007, 68, 384–386. [Google Scholar] [CrossRef] [PubMed]
  4. Chaovalitwongse, W.; Jeong, Y.; Jeong, M.K.; Danish, S.; Wong, S. Pattern Recognition Approaches for Identifying Subcortical Targets during Deep Brain Stimulation Surgery. IEEE Intell. Syst. 2011, 26, 54–63. [Google Scholar] [CrossRef]
  5. Groiss, S.J.; Wojtecki, L.; Südmeyer, M.; Schnitzler, A. Review: Deep brain stimulation in Parkinson’s disease. Ther. Adv. Neurol. Disord. 2009, 2, 379–391. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Itakura, T. Deep Brain Stimulation for Neurological Disorders: Theoretical Background and Clinical Application; Springer: New York, NY, USA, 2015; ISBN 978-3-319-08475-6. [Google Scholar]
  7. Perlmutter, J.S.; Mink, J.W. Deep brain stimulation. Annu. Rev. Neurosci. 2006, 29, 229–257. [Google Scholar] [CrossRef] [PubMed]
  8. Moran, A.; Bar-Gad, I. Revealing neuronal functional organization through the relation between multi-scale oscillatory extracellular signals. J. Neurosci. Methods 2010, 186, 116–129. [Google Scholar] [CrossRef] [PubMed]
  9. Novak, P.; Nazzaro, J.M. Detection of the subthalamic nucleus in microelectrographic recordings in Parkinson disease using the high-frequency (> 500 Hz) neuronal background. J. Neurosurg. 2007, 106, 5. [Google Scholar] [CrossRef] [PubMed]
  10. Lin, S.-H.; Lai, H.-Y.; Lo, Y.-C.; Chou, C.; Chou, Y.-T.; Yang, S.-H.; Sun, I.; Chen, B.-W.; Wang, C.-F.; Liu, G.-T.; et al. Decreased Power but Preserved Bursting Features of Subthalamic Neuronal Signals in Advanced Parkinson’s Patients under Controlled Desflurane Inhalation Anesthesia. Front. Neurosci. 2017, 11, 701. [Google Scholar] [CrossRef] [PubMed]
  11. Cagnan, H.; Dolan, K.; He, X.; Contarino, M.F.; Schuurman, R.; van den Munckhof, P.; Wadman, W.J.; Bour, L.; Martens, H.C.F. Automatic subthalamic nucleus detection from microelectrode recordings based on noise level and neuronal activity. J. Neural. Eng. 2011, 8, 046006. [Google Scholar] [CrossRef] [PubMed]
  12. Rajpurohit, V.; Danish, S.F.; Hargreaves, E.L.; Wong, S. Optimizing computational feature sets for subthalamic nucleus localization in DBS surgery with feature selection. Clin. Neurophysiol. 2015, 126, 975–982. [Google Scholar] [CrossRef] [PubMed]
  13. Guerrero-Martínez, J.; Rosado-Muñoz, A.; Bataller-Mompean, M.; Francés-Villora, J.; Teruel, V.; Cervera-Ferri, A.; Martinez-Ricós, J.; Gutiérrez, A.; Martínez-Torres, I. Characterization of Microelectrode Records in Deep Brain Stimulation Applied to Parkinson’s Disease Patients. In VI Latin American Congress on Biomedical Engineering CLAIB 2014, Paraná, Argentina 29, 30 & 31 October 2014; IFMBE Proceedings; Springer: Cham, Switzerland, 2015; pp. 647–650. ISBN 978-3-319-13116-0. [Google Scholar]
  14. Dolan, K.; Martens, H.C.F.; Schuurman, P.R.; Bour, L.J. Automatic noise-level detection for extra-cellular micro-electrode recordings. Med. Biol. Eng. Comput. 2009, 47, 791–800. [Google Scholar] [CrossRef] [PubMed]
  15. Wong, S.; Baltuch, G.H.; Jaggi, J.L.; Danish, S.F. Functional localization and visualization of the subthalamic nucleus from microelectrode recordings acquired during DBS surgery with unsupervised machine learning. J. Neural Eng. 2009, 6, 026006. [Google Scholar] [CrossRef] [PubMed]
  16. Nisbet, R.; Elder, J.F.; Miner, G. Handbook of Statistical Analysis and Data Mining Applications; Elsevier: Amsterdam, Netherlands, 2009; ISBN 978-0-12-374765-5. [Google Scholar]
  17. Van der Heijden, F.; Duin, R.P.W.; de Ridder, D.; Tax, T.M.J. Classification, Parameter Estimation, and State Estimation: An Engineering Approach Using MATLAB; Wiley: Chichester, West Sussex, UK; Hoboken, NJ, USA, 2004; ISBN 978-0-470-09013-8. [Google Scholar]
  18. Eusebi, P. Diagnostic Accuracy Measures. Cerebrovasc. Dis. 2013, 36, 267–272. [Google Scholar] [CrossRef] [PubMed]
  19. Tang, J.; Alelyani, S.; Liu, H. Feature Selection for Classification: A Review. In Data Classification: Algorithms and Applications; CRC Press: Boca Raton, FL, USA, 2014; p. 37. [Google Scholar]
  20. Cai, J.; Luo, J.; Wang, S.; Yang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  21. Robnik-Šikonja, M.; Kononenko, I. Theoretical and Empirical Analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  22. Kalousis, A.; Prados, J.; Hilario, M. Stability of feature selection algorithms: A study on high-dimensional spaces. Knowl. Inf. Syst. 2007, 12, 95–116. [Google Scholar] [CrossRef]
  23. Przybyszewski, A.W.; Ravin, P.; Pilitsis, J.G.; Szymanski, A.; Barborica, A.; Novak, P. Multi-parametric analysis assists in STN localization in Parkinson’s patients. J. Neurol. Sci. 2016, 366, 37–43. [Google Scholar] [CrossRef] [PubMed]
Figure 1. On the left, three-dimensional (3-D) view of the brain structure with microelectrode recordings (MER) trajectories to target marked. On the right, neural activity registered of different subcortical structures as the MER descends into the brain.
Figure 1. On the left, three-dimensional (3-D) view of the brain structure with microelectrode recordings (MER) trajectories to target marked. On the right, neural activity registered of different subcortical structures as the MER descends into the brain.
Entropy 21 00346 g001
Figure 2. ROC Curves for the 4 versions of the KNN algorithm.
Figure 2. ROC Curves for the 4 versions of the KNN algorithm.
Entropy 21 00346 g002
Table 1. Average values for accuracy (ACC), specificity (ESP), sensitivity (SEN), area under the ROC curve (AUC), and index of diagnosis (DOR) performance indices of the four versions of proposed k-nearest neighbors (KNN) classifiers.
Table 1. Average values for accuracy (ACC), specificity (ESP), sensitivity (SEN), area under the ROC curve (AUC), and index of diagnosis (DOR) performance indices of the four versions of proposed k-nearest neighbors (KNN) classifiers.
KNN VersionACCESPSENAUCDOR
KNN0.8194 ± 0.00740.7863 ± 0.01140.8499 ± 0.00840.9028 ± 0.004821.9095 ± 2.1969
KNN_STA0.8563 ± 0.00580.8299 ± 0.00900.8807 ± 0.00670.9230 ± 0.003336.1316 ± 3.5980
KNN_PAT0.9358 ± 0.00330.9344 ± 0.00410.9371 ± 0.00640.9761 ± 0.0021213.8659 ± 24.6348
KNN_HEM0.9435 ± 0.00220.9422 ± 0.00420.9446 ± 0.00490.9815 ± 0.0019279.8760 ± 22.9661
Table 2. Average values for training and validation in seconds of the four versions of the proposed KNN classifiers.
Table 2. Average values for training and validation in seconds of the four versions of the proposed KNN classifiers.
KNN Versiont_Traint_Validation
KNN0.0514 ± 0.05790.4044 ± 0.0221
KNN_STA0.0369 ± 0.00220.4093 ± 0.0306
KNN_PAT0.0349 ± 0.00200.4037 ± 0.0296
KNN_HEM0.0357 ± 0.00380.4028 ± 0.0283
Table 3. p-values of the Friedman and Nememyi test. In bold, statistically significant differences.
Table 3. p-values of the Friedman and Nememyi test. In bold, statistically significant differences.
KNN versions ComparedACCESPSENAUCDOR
Friedman Test<0,001<0.001<0.001<0.001<0.001
KNN vs. KNN_STA0.17010.17010.17010.17010.1701
KNN vs. KNN_PAT<0.001<0.001<0.001<0.001<0.001
KNN vs. KNN_HEM<0.001<0.001<0.001<0.001<0.001
KNN_STA vs. KNN_PAT0.17010.17010.12430.17010.1701
KNN_STA vs. KNN_HEM<0.001<0.001<0.001<0.001<0.001
KNN_PAT vs. KNN_HEM0.17010.17010.29450.17010.1701
Table 4. Selection for each algorithm.
Table 4. Selection for each algorithm.
RSBSFSBBS
4-CL1-VAB2-RMS4-CL
5-TH2-RMS3-kur5-TH
6-PK3-kur4-CL6-PK
8-ZC4-CL5-TH8-ZC
12-SC5-TH6-PK13-SMAD
13-SMAD6-PK7-NE14-SCR
14-SCR7-NE8-ZC
8-ZC9-SBI
9-SBI12-SC
12-SC13-SMAD
13-SMAD14-SCR
14-SCR
Table 5. Measure for each model with feature selection. In bold, the highest values that were obtained for each indicator.
Table 5. Measure for each model with feature selection. In bold, the highest values that were obtained for each indicator.
Performance MeasuresKNN + RSKNN + BSKNN + FSKNN + BBSKNN
Accuracy (%)93.43 ± 0.3695.49 ± 0.2795.42 ± 0.3693.50 ± 0.3494.35 ± 0.22
Specificity (%)93.45 ± 0.5595.21 ± 0.3595.22 ± 0.3893.27 ± 0.5194.23 ± 0.42
Sensitivity (%)93.40 ± 0.6595.74 ± 0.4995.60 ± 0.5693.72 ± 0.5594.46 ± 0.49
AUC (%)97.50 ± 0.2398.68 ± 0.1498.67 ± 0.1797.58 ± 0.1998.15 ± 0.19
Table 6. p-values of Friedman and Nemenyi tests. Nemenyi test compares the classifiers by pairs including the models with all features as those resulting from the selection of features, as described in Section 2.6. In bold, statistically significant differences.
Table 6. p-values of Friedman and Nemenyi tests. Nemenyi test compares the classifiers by pairs including the models with all features as those resulting from the selection of features, as described in Section 2.6. In bold, statistically significant differences.
KNN Versions ComparedACCESPSENAUC
Friedman Test<0.001<0.001<0.001<0.001
KNN+RS vs. KNN+BS<0.001<0.001<0.001<0.001
KNN+RS vs. KNN+FS<0.001<0.001<0.001<0.001
KNN+RS vs. KNN+BBS0.9960.9740.9520.875
KNN+RS vs. KNN0.0550.1640.1330.024
KNN+BS vs. KNN+FS0.9891.0000.9750.999
KNN+BS vs. KNN+BBS<0.001<0.001<0.001<0.001
KNN+BS vs. KNN0.0470.0690.0130.118
KNN+FS vs. KNN+BBS<0.001<0.001<0.001<0.001
KNN+FS vs. KNN0.1530.0940.0740.065
KNN+BBS vs. KNN0.1340.0350.4850.251

Share and Cite

MDPI and ACS Style

Bellino, G.M.; Schiaffino, L.; Battisti, M.; Guerrero, J.; Rosado-Muñoz, A. Optimization of the KNN Supervised Classification Algorithm as a Support Tool for the Implantation of Deep Brain Stimulators in Patients with Parkinson’s Disease. Entropy 2019, 21, 346. https://0-doi-org.brum.beds.ac.uk/10.3390/e21040346

AMA Style

Bellino GM, Schiaffino L, Battisti M, Guerrero J, Rosado-Muñoz A. Optimization of the KNN Supervised Classification Algorithm as a Support Tool for the Implantation of Deep Brain Stimulators in Patients with Parkinson’s Disease. Entropy. 2019; 21(4):346. https://0-doi-org.brum.beds.ac.uk/10.3390/e21040346

Chicago/Turabian Style

Bellino, Gabriel Martin, Luciano Schiaffino, Marisa Battisti, Juan Guerrero, and Alfredo Rosado-Muñoz. 2019. "Optimization of the KNN Supervised Classification Algorithm as a Support Tool for the Implantation of Deep Brain Stimulators in Patients with Parkinson’s Disease" Entropy 21, no. 4: 346. https://0-doi-org.brum.beds.ac.uk/10.3390/e21040346

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop