Next Article in Journal
Sensor Classification Using Convolutional Neural Network by Encoding Multivariate Time Series as Two-Dimensional Colored Images
Next Article in Special Issue
Prediction of 30-Day Readmission for COPD Patients Using Accelerometer-Based Activity Monitoring
Previous Article in Journal
Time Series Multiple Channel Convolutional Neural Network with Attention-Based Long Short-Term Memory for Predicting Bearing Remaining Useful Life
Previous Article in Special Issue
Effect of Closed-Loop Vibration Stimulation on Heart Rhythm during Naps
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of Rheumatoid Arthritis Classification from Electronic Image Sensor Using Ensemble Method

1
Smart Assistive and Rehabilitative Technology (SMART) Research Group, Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS, 32610 Bandar Seri Iskandar, Malaysia
2
School of Engineering, Monash University Malaysia, 46150 Bandar Sunway, Malaysia
3
Faculty of Engineering, Technology and Built Environment, UCSI University, 56000 Kuala Lumpur, Malaysia
*
Author to whom correspondence should be addressed.
Submission received: 30 September 2019 / Revised: 4 November 2019 / Accepted: 6 November 2019 / Published: 27 December 2019
(This article belongs to the Special Issue Smart Sensors for eHealth Applications)

Abstract

:
Rheumatoid arthritis (RA) is an autoimmune illness that impacts the musculoskeletal system by causing chronic, inflammatory, and systemic effects. The disease often becomes progressive and reduces physical function, causes suffering, fatigue, and articular damage. Over a long period of time, RA causes harm to the bone and cartilage of the joints, weakens the joints’ muscles and tendons, eventually causing joint destruction. Sensors such as accelerometer, wearable sensors, and thermal infrared camera sensor are widely used to gather data for RA. In this paper, the classification of medical disorders based on RA and orthopaedics datasets using Ensemble methods are discussed. The RA dataset was gathered from the analysis of white blood cell classification using features extracted from the image of lymphocytes acquired from a digital microscope with an electronic image sensor. The orthopaedic dataset is a benchmark dataset for this study, as it posed a similar classification problem with several numerical features. Three ensemble algorithms such as bagging, Adaboost, and random subspace were used in the study. These ensemble classifiers use k-NN (K-nearest neighbours) and Random forest (RF) as the base learners of the ensemble classifiers. The data classification is accessed using holdout and 10-fold cross-validation evaluation methods. The assessment was based on set of performance measures such as precision, recall, F-measure, and receiver operating characteristic (ROC) curve. The performance was also measured based on the comparison of the overall classification accuracy rate between different ensembles classifiers and the base learners. Overall, it was found that for Dataset 1, random subspace classifier with k-NN shows the best results in terms of overall accuracy rate of 97.50% and for Dataset 2, bagging-RF shows the highest overall accuracy rate of 94.84% over different ensemble classifiers. The findings indicate that the efficiency of the base classifiers with ensemble classifier have substantially improved.

1. Introduction

Arthritis implies irritation of joints. Rheumatoid Arthritis (RA) is a typical type of joint pain. RA is a persistent provocative ailment that affects and decimates the joints of wrists, fingers, and feet. If left untreated, one can lose their ability to lead a normal life. Around 1 out of 100 individuals succumb to RA at a particular phase of their life. It could occur to anybody. RA is not an inherited illness. RA can emerge regardless of age; however, most generally begins around the age of 40 and 60. It is around three times more typical in women than in men. Today, about 1–2% of the world’s population are affected by RA. In a 10-year range of analysis, 50% of patients have extreme impairment and additionally a reduced life span from 3 to 18 years [1]. Over the years, soft computing has increasingly played an important role in helping ailment analysis in doctors’ decision process. The main aim of this study is to investigate the possibility of applying machine learning techniques to the analysis of RA characteristics.
Some of the feasible implementations with the ensemble machine learning include information on-request systems, smart homes control and monitoring systems, mobile games communication interfaces and hospital and outpatient care applications [2,3]. In addition, ensemble machine learning is also used in the development and evaluation of electroencephalogram (EEG)-based brain activity measured by biosensors [4]. Advances in sensor innovation in on-body wearable sensors and pressure sensors have enabled them to be used effectively for the classification of RA disease. Wearable glove sensor based on IOT was used to detect RA disease in [5] was proven to be reliable. Besides, thermal infrared camera sensors were used to sense the temperature changes in between finger joints in a study carried out in [6,7].
Machine Learning (ML) is a research area of computer science that aims to construct predictive data models using algorithms, techniques, and procedures to discover unexpressed associations within the data and create tools that exploit such combinations of description, prediction or prescription [7]. Machine learning was introduced in multiple medical areas and has demonstrated to be very precise in categorizing and detecting illnesses. Machine learning has become a commonly used technique in order to enhance medical service and disease diagnosis with the development of medical information in the medical sector [8]. In order to convert data into information, data mining has become a significant tool. The implementation of data mining in nearly all areas has proven to be effective. The implementation in the medical field has also proven to be comparable. Knowledge obtained through the use of data mining from medical information provides doctors with an extra base of awareness to decide on their procedures, therapy scheduling, risk analytics, and other projections.
Multidiscipline data mining, such as machine learning, statistics, model identification, visualization, etc., includes different methods such as relationship, correlation, optimization, making predictions, classification and clustering [9,10,11,12]. Classification and clustering are the two most commonly used information mining methods. Classification is a type of supervised learning technique with the aim of predicting the target group in the medical data where the classes are predefined for each event. On the other hand, clustering is an unsupervised teaching method, which aims to group information into communities of comparable items in which the items are comparable to each other and not similar items in other groups [13]. Researchers are currently able to use multiple classification algorithms in the literature; hence, it is difficult to choose the finest model for a specific dataset, as these traditional algorithms are subject to issues like computer simplicity, adherence to local minimum standards or overfitting with information for teaching purposes [14].
The ensemble learning is used mainly to enhance a classifier’s efficiency and is one of the most used methods for overcoming these issues. It is a process through which multiple classifiers can be combined in order to categorize new samples to improve prediction accuracy [15]. The most common methods used together are boosting, bagging, stacking, and voting. In the past, studies have demonstrated the use of machine learning algorithms such as k-Nearest Neighbours (k-NN), MultiLayer Perceptron (MLP) and Support Vector Machine (SVM) for classifying and predicting diseases [16,17,18]. Ensemble methods with bagging and adaboost techniques can improve the accuracy of classification substantially by combining various weak classifiers. Ensemble learning on different datasets was successfully evaluated and validated appropriately [17,18].
The study outlined in this article emphasizes the classification of medical data using different ensemble classification techniques such as Bagging, Adaboost and Random subspace together with the base classifiers such as k-NN and RF. In this research, two medical datasets have been used, where the first dataset is rheumatoid arthritis dataset from [19]. The second dataset is from the UCI Machine Learning Repository [8]. This UCI dataset is about Biomechanical features of orthopaedic patients (orthopaedics) presented into two groups with six numerical attributes. Both datasets are classifications of individuals into two classes. Classifications were assessed in terms of accuracy, precision, recall, F-measure and receiver operating characteristic (ROC) curve.
The paper is organized as following: Section 2 discusses the literature review in relation to the work involved. The proposed method and dataset are explained in Section 3. Section 4 presents the results and discussion of the experiment. Finally, Section 5 concludes the paper.

2. Literature Review

2.1. Related Work

RA is the weakening condition that, if diagnosed at an early stage, could be managed and controlled. This is the only strategy available at present for preventing permanent joint damage or severe deformity. The main issue in RA diagnosis is that RA does not have a particular indicative feature [20], furthermore, the long lag between infection and symptom, and analysis is because of late referral [21]. Das, R and Sengur, A [17] have reported that they were able to diagnose valvular heart disease using ensemble methods, with three classifiers as base classifier, namely, k-Nearest Neighbours (k-NN), Multi-Layer Percentron (MLP) and Support Vector Machine (SVM). The result showed that ensemble classifier can produce better results than single classifier in terms of accuracy. It can be concluded that with base classifier MLP, boosting method has a better performance compared to a single classifier, whereby the sensitivity of 97.3% and specificity of 94% were obtained. In the work by Shiezadeh, et al. [18], algorithm such as ADAboost with J48 learner was chosen as the base learner, decision tree and CSboost were applied to the dataset of 18 attributes. It is proven in his research that CSboost is able to provide the highest accuracy among the classification algorithms. Among the 18 attributes, it was found that gender, number of joints, elbow and knee joint and express sequence tag (EST) test result play the most important roles in the identification of rheumatoid arthritis. Its classification efficiency is as follows: sensitivity of 44% and accuracy of 85%. Early detection of RA and commencement of vigorous treatment during the early stage of the disease can decrease disease development and, in a few patients, even prompt medication free reduction [22]. Ogungbemile [22] classified rheumatoid arthritis (RA) into level of severity using Machine Learning focusing on Decision Tree. This experiment used the temperature dataset extracted from the joints of 18 control subjects and 13 patients that suffered from RA. Then, these datasets were used as the input to the C.5 software to create a classifier. The ratio of Min and Max measurements is the best discriminator, as it is able to generate a sensitivity up to 96% and specificity of 92%. When a person initially has joint torments, it might at first hard for a specialist to state that the individual certainly has RA. This is because there are numerous different reasons for joint torments. There is no single test which determined early RA to have 100% accurateness. Practically speaking, the determination of RA depends on the subject’s history, physical exam, radiographs, and results from the laboratory. This means developing an easy-to-use, non-invasive, and cost-effective diagnostic tool that bridges the gap and makes early diagnosis possible for everyone. The development of this technique has unlimited potential, can also be used as a telemedicine tool, and can therefore, provide diagnostic options for those living in remote areas where the community could not afford advanced equipment such as magnetic resonance imaging (MRI) and high-resolution ultrasound (HRUS). Given the cost-effectiveness of this technology, it can also be used in developing countries, where RA diagnostic tools currently do not exist, given the costly nature of some current images, such as MRI and HRUS.
The diagnosis of RA from a physician is based on symptoms shown by patients. This information can be collected either in the structure of a medical unit or in a patient’s records. This information can comprise of significant details [23], which can be useful. Several techniques and algorithms are being used to collect concealed data. For the past few years, different algorithms of machine learning have been applied on various medical data sets to predict and classify various diseases from medical datasets [24]. Chen et al. [25] used electronic medical record (EMR) data from rheumatology clinic visits to predict RA disease activity. From their study, SVMs with linear kernel were able to deliver the foremost strong performance. It is said to be the best performing model because on average, the 10-fold AUC (Area under Curve) was 0.831 with standard deviation of 0.0317. Furthermore, Carroll et al. [26] have performed analysis of a randomly selected datasets of 400 patients from Vanderbilt University Medical Center’s Synthetic Derivative and Northwestern medical Enterprise Data Warehouse (EDW) using electronic health records, natural language processing (NLP) systems, and codified data. They have classified the data by using Logistic regression algorithm. From their work, it was concluded that, after retraining the model obtained from Partners Healthcare with logistic regression method, the area under the receiver operating curve was found to be 92% for Northwestern and 95% for Vanderbilt compared with 97% obtained at Partners. However, a higher specificity of 97% was obtained compared with 72% obtained at Partners. Das et al. [27] reported that using distinctive attribute correlated to the pathology can clearly define classification of fibromyalgia and rheumatoid arthritis. The results have shown that using ADAboost classifier is able to achieve a success rate of 89%, while reaching 97.8596% in the best case. Research by Ku Abd. Rahim et al. [2] showed that SVM with ensemble classifiers produced the best overall accuracy rate of 99.22% for classification of human daily activities using ensemble methods. Research reported by Nurhanim [3] provided the best overall classification rate of 98.57% with Multiclass support vector machine polynomial kernel (MC-SVM-Polynomial) for classification of daily physical activities such as walking, sitting, standing, laying, walking upstairs and downstairs. In general, the conclusion is that the ensemble methods produce better results in comparison with other algorithms.
Supervised learning checks for the relations between a number of examples (input functions) and one or more defined targets (output variables) and then produce a model that predicts the output correctly when there is new set of unlabelled input data. The training data should be marked with the right input-output combinations for supervised learning to work. Supervised learning comprises both classifications, which involve predicting the class or category provided with a new set of input (therefore the result is a discreet variable), and regression, where a continuous variable value is to be estimated for a new sample. Classification and regression are types of supervised learning. As an example, in order for researchers to determine whether the treatment response of patients treated with special therapy can be predicted by a series of clinical features, a supervised learning algorithm may be applied to datasets in which each patient record contains a set of clinical features of interest, as well as a label that shows the degree of disease response. Another word for supervised learning is to learn from examples. This paper focuses on the classification method from the supervised learning. Classification techniques such as Naïve Bayesian classification, support vector machines (SVM), decision trees, logistic and linear regression, random forests and k-nearest neighbours are based on supervised learning algorithms [28,29,30].

2.2. Sensing for RA

Diagnosis and sensing of RA is a challenging proposition because the associated clinical features are broad, and onset is generally subtle. The American College of Rheumatology (ACR) [31] has prescribed criteria for the characterization of RA (the purported “ACR 1987 revised criteria”). The disease usually affects the metacarpophalangeal, wrist, and proximal interphalangeal joints. Non-invasive imaging technologies such as MRI, radiography, ultrasound, and thermal imaging are used for the diagnosis of RA. Researchers from [19] tried to investigate RA by automating the procedure with the use of lymphocytes in blood cell pictures. Patients may suffer without apparent joint swelling at early stages of the disease, at which radiography fails to show [22]. In addition, magnetic resonance imaging (MRI) can distinguish aggravation and joint swelling; however, it is costly and not realistic for redundant utilization [32]. Another technique, ultrasound, showed more prominent dependability and affectability in the locating of synovitis and emanation as to clinical examination [32]. The computed tomography (CR) is an important tool for the detection of structural bone abnormalities and the reduction of the articulated area [33]. However, this method could not detect the earliest stages of synovitis of the soft tissue and therefore reducing the rheumatologist’s information on the activity of the disease [34]. Lastly, infrared, also known as thermal imaging, is a technique that uses thermographic camera to capture invisible infrared images, and those images are converted to normal images that are visible to the human eye. Many different thermal imaging systems have been tested over the years since dedicated medical thermographs have been available.

2.2.1. Definition of RA

Rheumatoid Arthritis (RA) is a persistent provocative ailment that effects and decimates the joints of wrists, fingers, and feet. If left untreated, one can lose their ability to lead a normal life [1]. A joint is a structure at which two parts of the bones are fitted together. Joints permit development and adaptability of different parts of the body. Muscles that take out on ligaments that are appended to bone caused the motion of the bones. The edge of bones is covered by the ligaments. Synovial liquid, which is a thick liquid, is located between the ligaments of two bones that make up a joint. The liquid smoothens up the joint, which permits continuous operation among the bones. The synovium is known as the tissue that encompasses a joint. The cell of the synovium produces the synovial liquid. Capsule is known as external piece of the synovium. It is hard, provides the joint steadiness, and prevents the bones from moving out of joint [35]. Encompassing tendons and muscles additionally help to provide support and solidness to joints. RA is categorized as an autoimmune illness. Synovitis is a key characteristic of RA and is directly linked to multiple alterations to the underlying joint physiology. The differences between a healthy synovium and an inflamed synovium are strikingly clear from histological analysis. Figure 1 shows the effects of synovium on healthy and typical joint exhibiting symptoms of RA.
The revised criteria for classification of RA was based on the definitions of the ACR 1987 [31]. Based on this definition, a subject is said to have RA on the off chance that he or she has fulfilled no less than four of the seven criteria for no less than a month and a half.

2.2.2. Sensors for Detecting RA

There are different sensor elements being used to capture the data. The various types of sensors are used to acquire data to diagnose RA disease are shown in Figure 2.
The authors of previous research [5,6] have used thermal infrared camera (TIR) imaging system as a sensor to diagnose RA. This non-invasive thermal imaging instrument has demonstrated the ability in the assessment and diagnosis of RA. TIR imaging may render a diagnosis without physical symptoms feasible because it is a temperature measurement over a certain joint showing as an RA indicator. A characteristic of RA is inflammation in the articulations affected, which can easily be seen by thermal imaging, but not by the naked eye, as the temperature rises automatically around those articulations. The two authors [5,6] used TIR sensor for different parts of the body. From [5], the author used TIR sensor to acquire image analysis of the elbows, wrist and hand joints, knees, ankles and feet joints. Furthermore, authors in [7] used TIR sensor to capture the image of finger joints.
Moreover, the author of [36] used the acquisition of optical pictures using a CCD (charge-coupled device) image sensor from the side of the dorsal hand. In order to detect optical modifications, a non-tomographic, single-wavelength, CW transillumer has earlier been provided, in which a light source focused on the finger’s side, using a CCD camera sensor to collect the images on the palmar side.
The author from [37] used a wearable sensor on the wrist in terms of accelerometer to access the pain and stiffness experienced by the patients. The research shows that regular, sensor based, controlled, unmonitored physical exams can be feasible and useful to objectively evaluate the effect of the illness on home environment function. Lightweight tiny accelerometers, generally worn on the wrist, hip or upper arm, allow for a constant recording over several days of information relating to movements (e.g., trunk, wrist, or ankle accelerations). The patients are required to perform an unsupervised version of a standard motor task, known as the Five Times Sit to Stand (5 × STS) test. This study has proved that the usage of sensor based, and unsupervised physical tests definitely has an impact in diagnosing this disease in a home environment.
Researchers in [5] captured the real time data from wearable glove measurement sensor and a 3D interface. This system will accurately quantify patients accurately in grades, maximum and minimum joint range, flexure, extension, and abduction of the finger or thumb joints, and compares the joint range of the normal ROM values in order to detect the deformity of the hand and rigidity of moveable finger joint.

3. Methods

3.1. Proposed Model

The methodology carried out in this research is defined in the following steps in Figure 3. RA can reduce the level of lymphocytes. Lymphocytopenia, a medical term used for a low count of white blood cells may be a RA complication called Felty syndrome that presents signs of severe pain, morning stiffness and joint swelling. Automatic assessment of lymphocytes requires a strong algorithm of segmentation that identifies the blood smear picture lymphocytes. Precision is essential as the classification depends on these segmentation steps. Analysis of the classification of white blood cells using features extracted from lymphocyte images is very effective.
According to Figure 3, blood smear images collected under 100× microscope with an effective 1000 magnification. As for RA, only lymphocytes must be taken as the region of interest (ROI) and the vital characteristics must be taken from it. Generally, digital microscopes generate pictures in (red, green and blue) RGB color space. In order to make the segmentation easier, RGB images are then converted to binary. Pre-processing is an important step in removing noise from the images. Next, the goal is to segment the images from white blood cell (WBC) and lastly to separate lymphocytes from the background. Feature extraction is the steps to discover features from the image. Then, classification would be carried out using Ensemble classifier. Ensemble classification method includes two main processes, which are training and testing. The Ensemble classifier’s goal, based on the training dataset, is to deliver a decision model that can predict the test information class label given by the feature vector. Moreover, algorithm could be tuned accordingly to produce a better accuracy of the prediction model. This experiment is using the base classifier method consist of k-NN and Random Forest to generate classifier models together with ensemble classifiers of bagging, boosting and random subspace. The assessment criteria for this classification algorithm includes accuracy, precision, recall, F-measure, and AUC that can be found from the confusion matrix output.

3.2. Database

Two datasets were used in this study. Dataset 1 is obtained from [19]. Dataset 1 was collected using lymphocyte in the blood cell images acquired from microscopic image. This dataset contains 40 instances, each containing 9 attributes named area, perimeter, circularity, integrated density, Median, Skewness, raw integrated density, and Roundness and solidity. Therefore, in total, 360 data (40 × 9) were used.
Dataset 2 is from the UCI Machine Learning Repository [8]. This UCI dataset comprises of Biomechanical features of orthopaedic patients (orthopaedics) consists 6 attributes and 310 instances named pelvic tilt numeric, lumbar lordosis angle, pelvic incidence, pelvic radius, sacral lope, lumbar lordosis angle, and degree spondylolisthesis. Dataset 2 is a benchmark dataset for this study, as it posed a similar classification problem with several numerical features which were provided for the classification. Models were run first on the orthopaedics dataset to ensure they were correctly implemented. Every entry in the Datasets 1 and 2 corresponds to a different patient.

3.3. Data Pre-Processing

This is a significant preliminary phase in the data mining method involving cleaning, assimilation, reconstruction, extraction and selection of features [38]. The precision and reliability of the assessment is determined by the data quality. The information may contain missing values, noise, irrelevance and redundancy. The data mining process will result in misleading results if data is not handled properly. The RA used in this study has 5 missing values. Missing data can be handled in many different ways. For this approach, the RA dataset was refined by replacing the records containing missing value with the mean value of the variable. Data pre-processing therefore functions as a preliminary method in which information can be transformed for clustering.

3.4. Classification Techniques—Proposed Methods

Recent developments in computer learning have led to methods that improve the performance of basic learning systems or enhance the capabilities of them. These learning systems have been called as ‘ensembles’. Ensembles are machine-learning processes that uses the ability from several models to obtain greater prediction than any model can on its own [39]. The algorithm building the model is called an inducer and an inducer is regarded as a classifier for certain training sets [40]. An ensemble consists of several inducers, frequently known as the base classifier. A base learner is an algorithm which gets a number of labelled cases as an input and generates a model to simplify those cases. Predictions for new unclassified instances are determined using the model created. The ensemble classification shows the use of multiple classifier combinations. These combinations operate in three steps. The first step is to construct multiple training subsets from a training set. In second step, each classifier is built exclusively on the basis of both the algorithm and the subset of data training. In the final step, base classifier outcomes are incorporated, and ultimate outcomes are acquired in a greater tier phase called ensemble classifier. It is known that an ensemble often performs far better than the individual classifiers which form it. In this study, two medical datasets were evaluated with ensemble methods such as Bagging, Adaboost, and Random Subspace. An ensemble inducer can be of any sort of base classifiers such as k-NN, decision tree, linear regression and other type of base learner algorithm. In this research work, the base learners used were k-NN and RF. An overview of the ensemble methods can be discovered in [41] for bagging, [42] for Adaboost and [43] for Random subspace.
The three ensemble methods and three basic learner algorithms described above were used to classify the medical datasets based on a data mining tool known as the WEKA 3.9 version [44], using a holdout model assessment system (containing 70% of training set and 30% of test set) and a 10-fold cross-validation method.

3.5. Performance Measures

The performance measures used in this study are based on the following terms:
Accuracy = T P + T N T P + T N + F P + F N
Precision = T P T P + F P
Recall = T P T P + F N
F - measure = 2 ( T P ) 2 ( ( T P ) + ( F N ) )
where TP—True Positive, TN—True Negative, FP—False Positive, and FN—False Negative.
The AUC is a Receiver Operating Curve (ROC) curve scalar value which is used to explore different model’s performance. The AUC measures the model’s ability to distinguish between distinct class values. It is a graphical strategy for showing the trade-off between a classifier’s true positive rate (recall) and false positive rate. The true positive rate is also referred to as recall in Equation (3) and false positive rate is defined as the following term:
False   positive   rate = F P F P + T N

4. Results and Discussion

4.1. Performance Measures of Dataset 1

4.1.1. 10-Fold Cross-Validations for Precision, Recall, F-Measure and ROC Measures

The performance measures of different ensemble machine learning algorithms together with two base classifiers k-NN and RF are shown in the following section. Table 1 shows the performance measures of the developed ensemble model with k-NN and random forest as base classifier. Figure 4 shows the weighted average of various parameters of ensemble methods with base classifiers of k-NN and Figure 5 shows the weighted average of various parameters of ensemble methods with base classifiers of random forest.
Figure 4 and Figure 5 show that in the case of all test configurations, the metric values were very similar. The results of random subspace classifier with k-NN and random forest as base classifier on the 10-fold cross-validation method perform equally well with precision and recall with 97.60% and 97.50 respectively compared to another ensemble classifier. Higher recall percentage can diagnose the correct disease for the patient. With base classifier of k-NN and RF, Adaboost and random subspace gives a higher precision result range of 97.60% to 95.30% compared to Bagging with 89.80% to 91.10%. In the case of recall, the percentage lies in the range of 97.50% to 95.00% for Adaboost and random subspace compared to bagging with 90.00%. Excellent results were gained in terms of F-measure for both base classifiers (most of the values are over 90%). From Table 2, all of the models developed AUC statistics are above 0.90. This means that all models perform more than 90%. The AUC results for k-NN were from 94.70% to 99.70% and 97.00% to 100% for RF. Table 2 shows an example of variability score of Dataset 1 using combination of random subspace with k-NN as base classifier.
Table 3 shows the AUC value for Dataset 1 with k-NN as the base classifier with the three-ensemble classifier, bagging, boosting and random subspace using cross validation method.
As shown in Table 3, excellent results were gained for the AUC for each ensemble classifier with k-NN as the base classifier. As stated in the literature, the model is perfect when its area under the curve is equal to 1, whereas the model makes a random assessment when its area below the curve is equivalent to 0.5. In this case of 3 ensemble classification with k-NN as base classifier, AUC value seems to be close to 1 (Plot area = 0.947) which indicates an accurate classification. Representative ROC curves for k-NN are displayed in Figure 6 based on Table 3.

4.1.2. Holdout Method for Precision, Recall, F-Measure and ROC Measures

Table 4 shows the performance measures with base classifiers of k-NN and RF on holdout method, Figure 7 shows the weighted average of various parameters of ensemble methods with base classifiers of k-NN and Figure 8 shows the weighted average of various parameters of ensemble methods with base classifiers of RF.
From Table 4, the results of ensemble classifier using holdout method indicates RF as the base learner generates a better precision of results ranging from 83.30% to 86.60% compared to k-NN with 68.80% to 81.00%. Results of recall evaluation shows that RF as base learner gained greater range of results from 75.00% to 95.80%. The results show that the RF base classifier with ensemble algorithm has the highest value for both precision and recall. Thus, it has the highest value for F-measure ranging from 78.93% to 86.80%. RF with ensemble algorithm is found to have an enhanced value of ROC with results ranging from 97.10% to 100% compared to k-NN with 77.10% to 95.80%. Table 5 shows the ROC curve for Dataset 1with k-NN as the base classifier with the three-ensemble classifier, bagging, boosting and random subspace using holdout method.
As shown in Table 5, the results for the AUC evaluation with k-NN as the base classifier for bagging recorded the best AUC value of 0.958. Representative ROC curves for k-NN are displayed in Figure 9 based on Table 5.

4.1.3. 10-Fold Cross Validation and Holdout for Overall Accuracy Rate

Figure 10 and Figure 11 shows the accuracy rate classification for Dataset 1 for ensemble method with base classifier k-NN and RF using 10-fold cross validation and holdout method.
As shown in Figure 10, the accuracy rate of ensemble classifier with the two base classifier k-NN and RF have the same level of performance. Using bagging ensemble classifier, accuracy rate: 90.00%, with Adaboost ensemble, accuracy rate: 95.00% and random subspace ensemble, accuracy rate: 97.50%. Figure 10 predict that random subspace performs in an excellent manner with k-NN and RF as the base classifier, while all other ensemble classifiers performs well. All test configurations delivered models that classified more than 90% of instances correctly.
The accuracy rate of ensemble methods with base learners k-NN and RF 10-fold cross validation method for Dataset 1 have also been presented in the Table 6 for better visualization.
Results in Table 6 presented that all of the testing configurations gained very good results. In the case of base classifier, k-NN and RF perform equally well with the same accuracy rate for all ensemble classifier. It is important to note that in the case of the ensemble classifier random subspace with both base classifier k-NN and RF, the accuracy is significantly improved compared to ensemble classifier bagging. The accuracy difference between ensemble classifier bagging compared to random subspace is equals to 7.5. Figure 11 shows the Accuracy rate of Dataset 1 with holdout method.
Results from Figure 11 predict the accuracy rate for respective ensemble classifier with RF using holdout method is higher with a range of 75.00% to 86.96% compared to k-NN with range of 66.67% to 71.43%.
The results from the Figure 11 have been also presented in Table 7.
From Table 7, while the classifications in all the ensembles have good performance with base classifier generally, it turns out that base classifier k-NN was a poor performer and RF was the optimal classifier with the ensemble classifier. Overall results of 10-fold cross-validation model evaluation provide better results for both base learners as compared to holdout.

4.2. Performance Measures of Dataset 2

4.2.1. 10-Fold Cross Validation Method for Precision, Recall, and ROC Measures

Table 8 presents the performance measures of the developed ensemble model with k-NN and random forest as base classifier with 10-fold cross validation method. Figure 12 shows the weighted average of various parameters of ensemble methods with base classifiers of k-NN and Figure 13 shows the weighted average of various parameters of ensemble methods with base classifiers of random forest.
From Table 8, the best precision results bagging classifier on 10-fold cross validation method shows that RF gained 95.00% compared to k-NN which gained 83.50%. For other ensemble classifier, Adaboost-k-NN and Adaboost-RF obtained precision of 82.40% and 94.30% respectively compared to random subspace–k-NN and random subspace-RF which recorded 84.20% and 89.20% respectively. Recall results achieved 94.80% using bagging-RF compared to bagging-k-NN which achieved 83.20%. The recall results obtained 84.20% for both random subspace-k-NN and Adaboost-RF. With Adaboost-k-NN, recall recorded 81.60% and random subspace-RF recorded recall result of 89.70%. With k-NN as base classifier, F-measure performance ranging from 82.00% to 84.20%, while with RF as base classifier, F-measure performance ranging from 89.55% to 94.90%. The AUC had high values for all of the testing configurations (AUCmax = 92.50%) with RF base classifier. Lower results were gained for the ROC with k-NN base classifier (AUCmin = 80.70%)
Table 9 shows the AUC value for Dataset 2 with RF as the base classifier with the three-ensemble classifier, bagging, boosting and random subspace using cross validation method.
As shown in Table 9, excellent results were gained for the AUC evaluation for each ensemble classifier with RF as the base classifier. Representative ROC curves for RF are displayed in Figure 14 based on Table 9.

4.2.2. Holdout Method for Precision, Recall, and ROC Measures

Table 10 demonstrate the performance measures of developed ensemble model with k-NN and random forest as base classifier with holdout method. The performance measures of ensemble methods with base learners k-NN and RF for Dataset 2 have also been presented in the Figure 15 and Figure 16 for greater visualization.
As demonstrated in Table 10, the results of bagging classifier with k-NN and RF base classifier evaluation on holdout method obtained the best precision with 85.10%(RF) and 82.10%(k-NN) respectively. For other ensemble classifier, Adaboost-k-NN and Adaboost-RF shows a higher precision performance of respectively 80.90% and 84.00% compared to random subspace–k-NN which recorded 75.00% and random subspace-RF 80.90%. For recall, bagging and Adaboost performed similarly as they both obtained same precision value of 83.80%. Random subspace-k-NN obtained a recall of 72.40% and random subspace-RF obtained 79.00% which performance is lower than bagging and Adaboost classifier. The above bar chart reveals that F-measure of the RF base classifier with ensemble classifier ranging is higher than k-NN base classifier. The AUC had high values for all of the testing configurations (AUCmax = 90.70%) with RF base classifier. Lower results were gained for the ROC with k-NN base classifier (AUCmax = 76.70%).
Table 11 shows the AUC value for Dataset 2 with RF as the base classifier with the three-ensemble classifier, bagging, boosting and random subspace using holdout method.
As shown in Table 11, for each bagging and Adaboost ensemble classifier with RF as the base classifier produced the same results for the AUC evaluation, 0.907. For ensemble random subspace, AUC evaluation recorded 0.849. Representative ROC curves for k-NN are displayed in Figure 17 based on Table 11.

4.2.3. 10-Fold Cross Validation and Holdout for Overall Accuracy Rate

Figure 18 and Figure 19 show the accuracy rate classification for Dataset 2 for ensemble method with base classifier k-NN and RF using 10-fold cross validation and holdout method.
Figure 18 predict that bagging-RF happens to be the best performer with accuracy of 94.84%. With the base classifier k-NN, the accuracy for ensemble classifier range from 81.61% to 84.19%. On the other hand, Random subspace-RF is found to be a poor performer with the lowest accuracy of 79.68%. With base classifier RF, the accuracy for ensemble classifier range from 79.68% to 94.84%.
The results from Figure 18 obtained with the use of Dataset 2 with 10-fold cross validation method are also shown in Table 12.
Table 12 predict that RF works better with the ensemble bagging and Adaboost with accuracy of more than 84%. On the other ensemble random subspace, RF accuracy performance fluctuated to 79.68%. While with base classifier k-NN, all ensemble classifier accuracy performance is above 80%
Figure 19 shows the Accuracy rate of Dataset 2 with holdout method.
Results from Figure 19 predict the accuracy rate for respective ensemble classifier with RF using holdout method is higher with a range of 79.05% to 83.81% compared to k-NN with range of 72.38% to 80.95%.
The results from the Figure 19 have been also displayed in the Table 13. The figure clearly shows the diversity of values of particular metrics for different ensemble classifier.
Table 13 gives the measurement of accuracy improvement enhancement in regard to the different classifiers considered here. RF with bagging and Adaboost ensemble classifier happens to be the best performer with accuracy of 83.81%. Overall results of 10-fold cross-validation model evaluation provide better results for both base learners as compared to holdout.

4.3. Comparative Analysis

The overall classification accuracy rate comparison of the various classification methods with previous work is represented in Table 14.
From Table 14 it is seen that the proposed model for Dataset 1 with random subspace-k-NN obtained an accuracy rate of 97.50 and for Dataset 2 with bagging-RF achieved an accuracy rate of 94.84 using 10-fold cross validation. The proposed model gives an enhanced accuracy result on the same data set that have been used when compared with the previous methods by Chokkalingam and Komathy [19] that has achieved an accuracy rate of 96.50% and Hasan and Islam [8] that obtained an accuracy rate of 92.00%. The proposed model classification accuracy is higher when compared to work done by previous researchers for both Datasets 1 and 2. The proposed model for Dataset 1, random subspace-k-NN has the ability to produce a higher accuracy rate. This should be because of k-NN’s stabilization with regard to training sets alterations. Moreover, k-NN is a highly stable classifier, whereas other ensemble methods like bagging and boosting are developed by sub-sampling the training examples don’t operate well with stable classifiers. This corresponds to the results reported by Skurichina and Duin [45] and Zhi-Hua Zhou [46]. K-NN is benefitted by random subspace, as Sun et al. [47] pointed out, because random subspace is classified in much smaller subspaces, which can decrease adverse noise impact for neighbour computation.
The proposed model bagging-RF for Dataset 2 showed good accuracy rate. Random forest is an improvement of the bagging which improves the range of variable. The random forest classifier combines the benefits of bagging with random selection of functions [48]. This paper is able to validate one of the advantages listed in their paper [49] namely, random forest method has excellent tolerance to noise and great accuracy of classification and the random forest method can be competent in view of the enormous amount of large data.
The result deduced that the proposed model random subspace-k-NN for Dataset 1 and bagging-RF for Dataset 2, the 10-fold cross validation method outperforms the holdout method. Several researchers such as Witten and Frank [24], Bengio et al. [50], Sakr et al. [51], Kim [52] and Dietterich [53] supported this claim that very often, such a cross-validation produces excellent outcomes and is better than any spilt. The k-fold cross validation can be repeated in t-times to reduce the randomness influence caused by the spread of data [46]. It must be noted that the model with high accuracy, recall, precision, ROC and F-measure is the best predictive model and highly desired. Ahmad et al. [54] and Zhi-Hua Zhou [46].
It can be concluded that an ensemble method is often considerably more precise than a single learner, and ensemble techniques in many real-world assignments have already been very successful. There are three threads pointed out by Zhi-Hua Zhou [46] that have led to the present area of ensemble methods: the combination of classifiers, ensemble of weak learners and mixture of experts Ensemble of weak learners was mostly researched in the community of machine learning. In this field, the researches focused on the weak learners and create powerful algorithms to enhance the performance from weak to strong [55]. This work by the researches has resulted to the development of well-known ensemble techniques such as Adaboost and Bagging. In general, most ensemble methods for the type of base classification used to construct the ensemble are self-determined. This is an important benefit that enable a specific classifier to be developed which could be best known for a particular application.

5. Conclusions

In this research, the usage of different ensemble classifiers with different base learner algorithms was implemented for different medical data sets. This study was conducted on rheumatoid arthritis and biomechanical features of orthopaedics datasets. The efficiency and comparison of the algorithms is also done using three popular ensemble classifiers: bagging, boosting, and random subspace with base learner such as k-NN and RF. The experimental results also showed that ensemble classifier is dependent on the type of base classifiers and type of data used. Moreover, this study also deduced, there are selected classifiers that should be used for each dataset. Several performance measures were utilized to evaluate the effectiveness of the classification techniques such as overall accuracy, recall, precision, ROC and F-measures. There are two evaluation methods, holdout and 10-fold cross validation methods were used to identified to yield the best model. The results of the comparison algorithms show that for Dataset 1, random subspace classifier with k-NN shows the best results in terms of overall accuracy rate compared to different ensemble classifiers. On the other hand, for Dataset 2, bagging-RF shows the highest overall accuracy rate over different ensemble classifiers. The comparison in the classification of both datasets shows the overall results of performance for recall, precision, ROC and F-measures shows a slight tendency favouring 10-fold cross validation method. It is concluded that using ensemble classifiers for diagnosing medical data enhance the classification accuracy, although different ensemble classifier would show different performance. In the future, different approaches and their application to our ensemble approach would be explored, which is crucial for the efficient diagnosis which would enhance the health index.

Author Contributions

I.E. led to the development of the concept and the structure of the research scheme, H.S. participated in the construction of the search queries, engaged in the query for literature, conducted studies and authored significant parts of the manuscript. I.E. and C.-K.L. provided significant critical analysis of the research findings. S.P. and E.N. assisted in the review and editing of the manuscript. All authors equally contributed to the rest of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ministry of Higher Education (MOHE), Malaysia through FRGS Grant, 0153AB-L28, and Universiti Teknologi PETRONAS.

Acknowledgments

The authors would like to thank Universiti Teknologi PETRONAS, Malaysia, Monash University Malaysia and UCSI University for supporting this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huizinga, T.W.; Pincus, T. Rheumatoid arthritis. Ann. Intern. Med. 2010, 153, ITC1-1. [Google Scholar] [CrossRef] [PubMed]
  2. Rahim, K.K.A.; Elamvazuthi, I.; Izhar, L.I.; Capi, G.; Rahim, K.N.K.A. Classification of Human Daily Activities Using Ensemble Methods Based on Smartphone Inertial Sensors. Sensors 2018, 18, 4132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Nurhanim, K.; Elamvazuthi, I.; Izhar, L.I.; Ganesan, T. Classification of Human Activity Based on Smartphone Inertial Sensor Using Support Vector Machine. In Proceedings of the IEEE 3rd International Symposium in Robotics and Manufacturing Automation (ROMA), Kuala Lumpur, Malaysia, 19–21 September 2017; pp. 1–5. [Google Scholar]
  4. Katona, J.; Kovari, A. A Brain–Computer Interface Project Applied in Computer Engineering. IEEE Trans. Educ. 2016, 59, 319–326. [Google Scholar] [CrossRef]
  5. Condell, J.; Curran, K.; Quigley, T.; Gardiner, P.; McNeill, M.; Winder, J.; Xie, E.; Qi, Z.; Connolly, J. Finger movement measurements in arthritic patients using wearable sensor enabled gloves. Int. J. Hum. Factors Model. Simul. 2011, 2, 276–292. [Google Scholar] [CrossRef]
  6. Pauk, J.; Wasilewska, A.; Ihnatouski, M. Infrared Thermography Sensor for Disease Activity Detection in Rheumatoid Arthritis Patients. Sensors 2019, 19, 3444. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Kim, K.J.; Tagkopoulos, I. Application of machine learning in rheumatic disease research. Korean J. Intern. Med. 2019, 34, 708–722. [Google Scholar] [CrossRef] [Green Version]
  8. Hasan, K.; Islam, S.; Samio, M.M.R.K.; Chakrabarty, A. A Machine Learning Approach on Classifying Orthopedic Patients Based on Their Biomechanical Features. In Proceedings of the Joint 7th International Conference on Informatics, Electronics & Vision (ICIEV) and the 2nd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Kitakyushu, Japan, 25–29 June 2018; pp. 289–294. [Google Scholar]
  9. Jenifer, S.; Parasuraman, S.; Kadirvelu, A. Contrast enhancement and brightness preserving of digital mammograms using fuzzy clipped contrast-limited adaptive histogram equalization algorithm. Appl. Soft Comput. 2016, 42, 167–177. [Google Scholar] [CrossRef]
  10. Jenifer, S.; Parasuraman, S.; Kadirvelu, A. An Efficient Biomedical Imaging Technique for Automatic Detection of Abnormalities in Digital Mammograms. J. Med. Imaging Health Inform. 2014, 4, 291–296. [Google Scholar] [CrossRef]
  11. Natarajan, E.; Kaviarasan, V.; Lim, W.H.; Tiang, S.S.; Tan, T.H. Enhanced Multi-Objective Teaching-Learning-Based Optimization for Machining of Delrin. IEEE Access 2018, 6, 51528–51546. [Google Scholar] [CrossRef]
  12. Natarajan, E.; Kaviarasan, V.; Lim, W.H.; Tiang, S.S.; Parasuraman, S.; Elango, S. Non-dominated sorting modified teaching–learning-based optimization for multi-objective machining of polytetrafluoroethylene (PTFE). J. Intell. Manuf. 2019, 1–25. [Google Scholar] [CrossRef]
  13. Tan, P.N.; Michael, S.; Vipin, K. Introduction to Data Mining; Pearson/Addison Wesley: Noida, India, 2016; ISBN 0-321-32136-7. [Google Scholar]
  14. Dietterich, T.G. Ensemble Methods in Machine Learning. Lect. Notes Comput. Sci. 2000, 1857, 1–15. [Google Scholar]
  15. Ethem, A. Introduction to Machine Learning, 2nd ed.; MIT Press: Cambridge, MA, USA, 2009. [Google Scholar]
  16. Kumar, N.; Khatri, S. Implementing WEKA for medical data classification and early disease prediction. In Proceedings of the 3rd IEEE International Conference on Computational Intelligence and Communication Technology (CICT), Ghaziabad, India, 9–10 February 2017. [Google Scholar]
  17. Das, R.; Sengur, A. Evaluation of ensemble methods for diagnosing of valvular heart disease. Expert Syst. Appl. 2010, 37, 5110–5115. [Google Scholar] [CrossRef]
  18. Shiezadeh, Z.; Sajedi, H.; Aflakie, E. Diagnosis of Rheumatoid Arthritis Using an Ensemble Learning Approach. In Proceedings of the Fourth International Conference on Advanced Information Technologies and Applications, Dubai, United Arab Emirates, 6–7 November 2015; pp. 139–148. [Google Scholar]
  19. Chokkalingam, S.P.; Komathy, K. Comparison of different classifier in WEKA for rheumatoid arthritis. In Proceedings of the International Conference on Human Computer Interactions (ICHCI), Chennai, India, 23–24 August 2013; pp. 1–6. [Google Scholar]
  20. Emery, P.; Symmons, D.P. Review What is early rheumatoid arthritis? definition and diagnosis. Baillieres Clin. Rheumatol. 1997, 11, 13–26. [Google Scholar] [CrossRef]
  21. Chan, K.W.; Felson, D.T.; Yood, R.A.; Walker, A.M. The lag time between onset of symptoms and diagnosis of rheumatoid arthritis. Arthritis Rheum. 1994, 37, 814–820. [Google Scholar] [CrossRef] [PubMed]
  22. Frize, M.; Adea, C.; Payuer, P.; di Primiro, G.; Karsh, J.; Ogungbemile, A. Detection of RA using infrared imaging. In Medical Imaging 2011: Image Processing; International Society for Optics and Photonics: Lake Buena Vista (Orlando), FL, USA, 2011; Volume 7962, p. 79620M. [Google Scholar]
  23. Witten, I.H.; Frank, E. Data Mining, Practical Machine Learning Tools and Techniques, 2nd ed.; Elsevier: Amsterdam, The Netherlands, 2005. [Google Scholar]
  24. Podgorelec, V.; Heričko, M.; Rozman, I. Improving Mining of Medical Data by Outliers Prediction. In Proceedings of the 18th IEEE Symposium on Computer-Based Medical Systems (CBMS’05), Dublin, Ireland, 23–24 June 2005; pp. 91–96. [Google Scholar]
  25. Lin, C.; Karlson, E.W.; Canhão, H.; Miller, T.A.; Dligach, D.; Chen, P.J.; Perez, R.N.G.; Shen, Y.; Weinblatt, M.E.; Shadick, N.A.; et al. Automatic Prediction of Rheumatoid Arthritis Disease Activity from the Electronic Medical Records. PLoS ONE 2013, 8, e69932. [Google Scholar] [CrossRef]
  26. Carroll, R.J.; Thompson, W.K.; Eyler, A.E.; Mandelin, A.M.; Cai, T.; Zink, R.M. Portability of an algorithm to identify rheumatoid arthritis in electronic health records. J. Am. Med. Inform. Assoc. 2012, 19, e162–e169. [Google Scholar] [CrossRef] [Green Version]
  27. Garcia-Zapirain, B.; Garcia-Chimeno, Y.; Rogers, H. Machine Learning Techniques for Automatic Classification of Patients with Fibromyalgia and Arthritis. IJCTT 2015, 25, 149–152. [Google Scholar] [CrossRef]
  28. Sanchez-Pinto, L.N.; Luo, Y.; Churpek, M.M. Big Data and Data Science in Critical Care. Chest 2018, 154, 1239–1248. [Google Scholar] [CrossRef]
  29. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning: With Applications in R; Springer: New York, NY, USA, 2013. [Google Scholar]
  30. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: New York, NY, USA, 2013. [Google Scholar]
  31. American College of Rheumatology. Available online: https://www.rheumatology.org/Portals/0/Files/1987%20Rheumatoid%20Arthritis%20Classification_Excerpt%201987.pdf (accessed on 12 November 2019).
  32. Soalding, S.J.; Kwoh, C.K.; Boudreau, R.; Enama, J.; Lunich, J.; Huber, D.; Denes, L.; Hirsch, R. 3D and thermal surface imaging produces reliable measures of joint shape and temperature: A potential tool for quantifying arthritis. Artritis Res. Ther. 2008, 10, R10. [Google Scholar] [CrossRef] [Green Version]
  33. Kaczmarek, M.; Nowakowski, A. Active IR-Thermal Imaging in Medicine. J. Nondestruct. Eval. 2016, 35, 19. [Google Scholar] [CrossRef] [Green Version]
  34. Cao, K.; Mills, D.M.; Thiele, R.G.; Patwardhan, K. Towards Quantitative Assessment of RA using Volerteic Ultrasound. IEEE Trans. Biomed. Eng. 2016, 63, 449–458. [Google Scholar] [CrossRef] [PubMed]
  35. Eutice, C. Rheumatoid Arthritis the Basics. 2012. Available online: http://arthritis.about.com/od/rheumatoidarthritis/p/rheumatoidfacts.htm (accessed on 15 June 2019).
  36. Lighter, D.; Filer, A.; Dehghani, H. Detecting inflammation in rheumatoid arthritis using Fourier transform analysis of dorsal optical transmission images from a pilot study. J. Biomed. Opt. 2019, 24, 066008. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Perraudin, C.G.; Illiano, V.P.; Calvo, F.; O’Hare, E.; Donnelly, S.C.; Mullan, R.H.; Sander, O.; Caulfield, B.; Dorn, J.F. Observational Study of a Wearable Sensor and Smartphone Application Supporting Unsupervised Exercises to Assess Pain and Stiffness. Digit. Biomark. 2018, 2, 106–125. [Google Scholar] [CrossRef]
  38. Baskar, S.S.; Arockiam, L.; Charles, S. A Systematic Approach on Data Pre-processing in Data Mining. Compusoft 2013, 2, 335. [Google Scholar]
  39. Dietterich, T.G. Ensemble Methods in Machine Learning. In Computer Vision—ECCV 2012; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2000; Volume 1857, pp. 1–15. [Google Scholar]
  40. Rokach, L. Ensemble-Based Classifiers. Artif. Intell. Rev. 2010, 33, 1–39. [Google Scholar] [CrossRef]
  41. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  42. Freund, Y.; Schapire, R.E. A desicion-theoretic generalization of on-line learning and an application to boosting. Model. Data Eng. 1995, 904, 23–37. [Google Scholar]
  43. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  44. Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H.; Hall, M. The WEKA data mining software: An update. ACM SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar]
  45. Skurichina, M.; Duin, R.P.W. Bagging, Boosting and the Random Subspace Method for Linear Classifiers. Pattern Anal. Appl. 2002, 5, 121–135. [Google Scholar] [CrossRef]
  46. Zhou, Z.H. Ensemble Methods: Foundations and Algorithms; Chapman & Hall/CRC: Cambridge, UK, 2012. [Google Scholar]
  47. Sun, S.; Zhang, C.; Zhang, D. An experimental evaluation of ensemble methods for EEG signal classification. Pattern Recognit. Lett. 2007, 28, 2157–2163. [Google Scholar] [CrossRef]
  48. Robnik-Šikonja, M. Improving Random Forests. In European Conference on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2004; pp. 359–370. [Google Scholar]
  49. Chen, J.; Li, K.; Tang, Z.; Bilal, K.; Yu, S.; Weng, C.; Li, K. A Parallel Random Forest Algorithm for Big Data in a Spark Cloud Computing Environment. IEEE Trans. Parallel Distrib. Syst. 2017, 28, 919–933. [Google Scholar] [CrossRef] [Green Version]
  50. Bengio, Y.; Grandvalet, Y. No unbiased estimator of the variance of k-fold cross-validation. J. Mach. Learn. Res. 2004, 5, 1089–1105. [Google Scholar]
  51. Sakr, S.; Elshawi, R.; Ahmed, A.; Qureshi, W.T.; Brawner, C.; Keteyian, S.; Blaha, M.J.; Al-Mallah, M.H. Using machine learning on cardiorespiratory fitness data for predicting hypertension: The Henry Ford ExercIse Testing (FIT) Project. PLoS ONE 2018, 13, e0195344. [Google Scholar] [CrossRef] [Green Version]
  52. Kim, J.-H. Estimating classification error rate: Repeated cross-validation, repeated hold-out and bootstrap. Comput. Stat. Data Anal. 2009, 53, 3735–3745. [Google Scholar] [CrossRef]
  53. Dietterich, T.G. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 1998, 10, 1895–1923. [Google Scholar] [CrossRef] [Green Version]
  54. Ahmad, W.S.W.; Zaki, W.M.W.; Fauzi, M.F.A. Lung segmentation on standard and mobile chest radiographs using oriented Gaussian derivatives filter. Biomed. Eng. Online 2015, 14, 20. [Google Scholar] [CrossRef] [Green Version]
  55. Chokkalingam, S.P.; Komathy, K. Detection of Rheumatoid Arthritis Using Lymphocyte Images. J. Adv. Res. 2014, 7, 3926–3934. [Google Scholar]
Figure 1. (a) Shows an example of normal joints in the hand, highlighting the location of a proximal interphalangeal joints (PIP) joint, (b) shows the inflamed joints of the hands with RA, highlighting the location of a PIP joint, (c) is an illustration of a typical healthy joint, and (d) an illustration of a typical joint exhibiting symptoms of RA. Figure 1 (c,d) also shows the schematic of a typical synovial joint.
Figure 1. (a) Shows an example of normal joints in the hand, highlighting the location of a proximal interphalangeal joints (PIP) joint, (b) shows the inflamed joints of the hands with RA, highlighting the location of a PIP joint, (c) is an illustration of a typical healthy joint, and (d) an illustration of a typical joint exhibiting symptoms of RA. Figure 1 (c,d) also shows the schematic of a typical synovial joint.
Sensors 20 00167 g001
Figure 2. Sensor used for each joint.
Figure 2. Sensor used for each joint.
Sensors 20 00167 g002
Figure 3. Steps in Methodology.
Figure 3. Steps in Methodology.
Sensors 20 00167 g003
Figure 4. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Figure 4. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Sensors 20 00167 g004
Figure 5. Weighted average of various parameters of ensemble methods with base classifiers of random forest.
Figure 5. Weighted average of various parameters of ensemble methods with base classifiers of random forest.
Sensors 20 00167 g005
Figure 6. Receiver operating characteristic (ROC) graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with k-NN classifier using 10-fold cross validation method Dataset 1.
Figure 6. Receiver operating characteristic (ROC) graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with k-NN classifier using 10-fold cross validation method Dataset 1.
Sensors 20 00167 g006
Figure 7. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Figure 7. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Sensors 20 00167 g007
Figure 8. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Figure 8. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Sensors 20 00167 g008
Figure 9. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with k-NN classifier using holdout method Dataset 1.
Figure 9. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with k-NN classifier using holdout method Dataset 1.
Sensors 20 00167 g009aSensors 20 00167 g009b
Figure 10. Accuracy rate of Dataset 1 with 10- fold cross validation method.
Figure 10. Accuracy rate of Dataset 1 with 10- fold cross validation method.
Sensors 20 00167 g010
Figure 11. Accuracy rate of Dataset 1 with holdout method.
Figure 11. Accuracy rate of Dataset 1 with holdout method.
Sensors 20 00167 g011
Figure 12. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Figure 12. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Sensors 20 00167 g012
Figure 13. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Figure 13. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Sensors 20 00167 g013
Figure 14. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with RF classifier using 10-fold cross validation method Dataset 2.
Figure 14. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with RF classifier using 10-fold cross validation method Dataset 2.
Sensors 20 00167 g014aSensors 20 00167 g014b
Figure 15. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Figure 15. Weighted average of various parameters of ensemble methods with base classifiers of k-NN.
Sensors 20 00167 g015
Figure 16. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Figure 16. Weighted average of various parameters of ensemble methods with base classifiers of RF.
Sensors 20 00167 g016
Figure 17. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with RF classifier using holdout method Dataset 2.
Figure 17. ROC graph activity of ensemble classifier bagging (a), Adaboost (b), random subspace (c) with RF classifier using holdout method Dataset 2.
Sensors 20 00167 g017
Figure 18. Accuracy rate of Dataset 2 with 10-fold cross validation method.
Figure 18. Accuracy rate of Dataset 2 with 10-fold cross validation method.
Sensors 20 00167 g018
Figure 19. Accuracy rate of Dataset 2 with holdout method.
Figure 19. Accuracy rate of Dataset 2 with holdout method.
Sensors 20 00167 g019
Table 1. Performance measures with base classifiers of k-NN and RF on 10-fold cross-validation method.
Table 1. Performance measures with base classifiers of k-NN and RF on 10-fold cross-validation method.
k-NNRandom Forest
PrecisionRecallF-MeasureAUCPrecisionRecallF-MeasureAUC
Bagging91.1090.0090.5494.7089.8090.0089.9097.70
Adaboost95.3095.0095.1594.5095.0095.0095.0097.00
Random Subspace97.6097.5097.5599.7097.6097.5097.55100.00
Table 2. Variability score of Dataset 1 of random subspace with k-NN as base classifier.
Table 2. Variability score of Dataset 1 of random subspace with k-NN as base classifier.
FoldPrediction
10.862
20.864
30.781
40.723
50.888
60.783
70.888
80.783
90.848
100.874
Table 3. Area under Curve (AUC) evaluation of ensemble with k-NN base classifier.
Table 3. Area under Curve (AUC) evaluation of ensemble with k-NN base classifier.
k-NN (10-Fold Cross Validation)
Ensemble ClassifierAUC
Bagging94.70
Adaboost94.50
Random Subspace99.70
Table 4. Performance measures with base classifiers of k-NN and RF on holdout method.
Table 4. Performance measures with base classifiers of k-NN and RF on holdout method.
k-NNRandom Forest
PrecisionRecallF-MeasureAUCPrecisionRecallF-MeasureAUC
Bagging81.0071.4075.9095.8086.6087.0086.8097.10
Adaboost78.7078.6978.7077.1085.6081.3083.39100.00
Random Subspace68.8066.7067.7387.5083.3075.0078.93100.00
Table 5. ROC evaluation of ensemble with k-NN base classifier.
Table 5. ROC evaluation of ensemble with k-NN base classifier.
k-NN (Holdout)
Ensemble ClassifierAUC
Bagging0.958
Adaboost0.771
Random Subspace0.875
Table 6. Overall performance measure ensemble classifiers for 10-fold cross validation method.
Table 6. Overall performance measure ensemble classifiers for 10-fold cross validation method.
Accuracy Rate
10-Fold Cross-Validation
Ensemble Methodk-NNRF
Bagging90.0090.00
Adaboost95.0095.00
Random Subspace97.5097.50
Table 7. Overall performance measure ensemble classifiers for holdout method.
Table 7. Overall performance measure ensemble classifiers for holdout method.
Accuracy Rate
Holdout
Ensemble Methodk-NNRF
Bagging71.4386.96
Adaboost78.5781.25
Random Subspace66.6775.00
Table 8. Performance measures with base classifiers of k-NN and RF on 10-fold cross validation method.
Table 8. Performance measures with base classifiers of k-NN and RF on 10-fold cross validation method.
k-NNRandom Forest
PrecisionRecallF-MeasureAUCPrecisionRecallF-MeasureAUC
Bagging83.5083.2083.3589.8095.0094.8094.9092.50
Adaboost82.4081.6082.0080.7094.3094.2094.2592.10
Random Subspace84.2084.2084.2087.4089.2089.7089.5587.50
Table 9. AUC evaluation of ensemble with RF base classifier.
Table 9. AUC evaluation of ensemble with RF base classifier.
RF (10-Fold Cross Validation)
Ensemble ClassifierAUC
Bagging0.925
Adaboost0.921
Random Subspace0.875
Table 10. Performance measures with base classifiers of k-NN and RF on holdout method.
Table 10. Performance measures with base classifiers of k-NN and RF on holdout method.
k-NNRandom Forest
PrecisionRecallF-MeasureAUCPrecisionRecallF-MeasureAUC
Bagging82.1081.0081.5484.2085.1083.8084.4590.70
Adaboost80.9079.0079.9476.7084.0083.8083.9090.70
Random Subspace75.0072.4073.6878.2080.9079.0079.9484.90
Table 11. AUC evaluation of ensemble with RF base classifier.
Table 11. AUC evaluation of ensemble with RF base classifier.
RF (Holdout Method)
Ensemble ClassifierAUC
Bagging0.907
Adaboost0.907
Random Subspace0.849
Table 12. Overall performance measure ensemble classifiers for 10-fold cross validation method.
Table 12. Overall performance measure ensemble classifiers for 10-fold cross validation method.
Accuracy Rate
10-Fold Cross-Validation
Ensemble Methodk-NNRF
Bagging83.2394.84
Adaboost81.6194.19
Random Subspace84.1979.68
Table 13. Overall performance measure ensemble classifiers for holdout method.
Table 13. Overall performance measure ensemble classifiers for holdout method.
Accuracy Rate
Holdout
Ensemble Methodk-NNRF
Bagging80.9583.81
Adaboost79.0583.81
Random Subspace72.3879.05
Table 14. Overall classification accuracy comparison with previous research work.
Table 14. Overall classification accuracy comparison with previous research work.
ReferenceEvaluation MethodDatasetClassification MethodOverall Accuracy Rate
Proposed model10-fold cross-validationDataset 1 (40 samples)Random subspace-k-NN97.50%
Proposed modelHoldoutDataset 1 (40 samples)Random subspace-k-NN66.67%
Proposed model 10-fold cross-validationDataset 2 (310 samples)Bagging-RF94.84%
Proposed modelHoldoutDataset 2 (310 samples)Bagging-RF83.81
S. P. Chokkalingam and K. Komathy [19] Cross validationDataset 1 (40 samples)ADT Tree96.50%
Hasan.K and Islam.S [8]HoldoutDataset 2 (310 samples)Decision Tree92.00%

Share and Cite

MDPI and ACS Style

Sharon, H.; Elamvazuthi, I.; Lu, C.-K.; Parasuraman, S.; Natarajan, E. Development of Rheumatoid Arthritis Classification from Electronic Image Sensor Using Ensemble Method. Sensors 2020, 20, 167. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010167

AMA Style

Sharon H, Elamvazuthi I, Lu C-K, Parasuraman S, Natarajan E. Development of Rheumatoid Arthritis Classification from Electronic Image Sensor Using Ensemble Method. Sensors. 2020; 20(1):167. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010167

Chicago/Turabian Style

Sharon, Ho, Irraivan Elamvazuthi, Cheng-Kai Lu, S. Parasuraman, and Elango Natarajan. 2020. "Development of Rheumatoid Arthritis Classification from Electronic Image Sensor Using Ensemble Method" Sensors 20, no. 1: 167. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop