1. Introduction
Tuberculosis (TB) is a deadly contagious disease that claims the lives of millions of people annually around the world. The World Health Organization (WHO) reports that, by 2020, it will have become the leading infectious agent-related cause of mortality [
1]. The bacterium
Mycobacterium tuberculosis that causes TB can be resistant to the antimicrobial medications that are used to treat it. Tuberculosis that does not respond to at least one first-line anti-TB drug is referred to as drug-resistant TB (DR-TB). Multidrug-resistant TB (MDR-TB) and extensively drug-resistant TB (XDR-TB) are two serious subtypes of DR-TB. Mignani et al. [
2] proposed the suggested drug for each form of drug-resistant tuberculosis, which is a separate therapy regimen according to the patient’s drug-resistance status.
Both drug susceptibility testing (DST) results and the occurrence of serious adverse events are necessary for the development of a treatment strategy for tuberculosis (TB). Options for anti-TB treatment include first-line drugs, second-line drugs, and newer, more effective drugs (
Figure 1).
The World Health Organization (WHO) recommends a 6-month-course of 2HRZ(E)/4HR for a patient with DS-TB as a standard treatment. If the patient’s clinical and radiographic status do not improve after 2 months of treatment with four drugs, DR-TB will be considered [
4]. Multidrug-resistant tuberculosis (MDR-TB) is diagnosed when a patient is resistant to at least rifampin and isoniazid, the two most powerful first-line anti-TB drugs, and requires therapy with some second-line and new generation anti-TB treatments. Extensively drug-resistant tuberculosis (XDR-TB) is diagnosed when a patient does not react to the most effective second-line treatments, such as fluoroquinolones or injectable aminoglycosides, and requires multiple second-line and new generation anti-TB drugs [
2].
The mismanagement of TB treatment and the dissemination of the disease from patient to patient are the main causes of the emergence and spread of drug-resistant TB. The treatment for tuberculosis typically consists in a medicine regimen to be taken daily for 6 months under close medical care. In crowded settings, such as prisons and hospitals, drug resistance is more likely to spread as a result of the improper or incorrect use of antimicrobial drugs, or the use of ineffective formulations of drugs (such as the use of single drugs, mediocre quality medicines, or bad storage conditions), and the premature interruption of treatment. Drug resistance can be diagnosed using specialized laboratory tests that examine the susceptibility of the bacteria to the medications, or that find patterns of resistance. These examinations may be molecular (such as Xpert MTB/RIF). Due to a lack of fully equipped laboratories, drug-resistance testing in a certain area can take a considerable amount of time; therefore, approaches that are both quick and cost-effective should be developed and used in this region. The gold standard for evaluating drug resistance is called drug susceptibility testing (DST). Beyond screening or triaging TB, a chest X-ray (CXR) is also used to recognize drug-resistant TB (DR-TB) due to the fact that CXRs are more convenient as a rapid, non-invasive, and cost-effective tool for TB diagnosis and therapy evaluation [
5,
6,
7,
8].
Machine learning (ML) and deep learning (DL) have become increasingly popular in recent years to categorize the response to anti-tuberculosis drugs due to their effectiveness and low-cost. Parkinson’s, breast cancer, and chronic kidney disease are only a few examples of diseases where ML and DL have been used to accurately forecast or estimate outcomes. Automated DL-based methods that rely on CXRs for tuberculosis (TB) diagnosis have been shown to be highly effective [
9,
10]. Karki et al. [
11] used radiological parameters and clinical patient data in conjunction with chest X-ray images to categorize tuberculosis as drug-resistant (DR) or drug-sensitive (DS).
Machine learning was ultimately used by Tulo et al. [
12] to categorize four distinct types of drug-resistant TB. This includes DS-TB, DR-TB, MDR, and XDR, which some refer to forms of tuberculosis that have developed resistance to at least some of the available treatments. A binary classification approach was used to categorize these four types of medication responses. The computational results show that there is room for improvement in the accuracy of the binary classification between DR-MDR, MDR-XDR, and DR-XDR, which lie between 82% and 93%. When diagnosing the TB medication response in the real world, there should be a single system that can classify all four types of drug reactions using a single CXR picture input, since it would be more convenient for clinicians to utilize.
Multi-class classification is a classification challenge in deep learning and machine learning that involves more than two classes or results. Images from the TB dataset require multi-class classification since there are at least four distinct treatment responses that must be recognized by the model. It is more challenging to predict the outcome for multi-class classification than for binary class classification since multiple outcomes must be gathered from the same dataset. Bayesian classifiers that employ Bayes’ theorem [
13], the decision tree model [
14,
15], and the ensemble deep learning models [
11] are all examples of methods that could be improved upon to create a more effective classification process.
After carefully reviewing the aforementioned literature, we have developed our study to target the following areas that have been overlooked to date:
The classification should be performed as four classes, which are DS, DR, MDR, and XDR, in order to suggest the correct recommended treatment program to the treaters.
The designed model has not yet been used as the suggestion system, such as mobile application or chatbot, in order to use the design model in real medical treatments.
The lowest accuracy is still 82%, which is not significantly high and can result in a lack of trust from doctors.
Inspired by these unmet needs, we propose the design of a comprehensive classification system for drug resistance that distinguishes between DS, DR, MDR, and XDR. To achieve this, the system will be developed using the following techniques: (1) Create an effective multi-class classification model, and (2) create an application utilizing the model created in (1). The remaining sections are organized as follows.
Section 2 will review the related work,
Section 3 will describe the materials and methods used in this study, and
Section 4 and
Section 5 will present the computational results and discussion. The investigation’s conclusions will be reported in
Section 6.
2. Related Work
The theory and related literature will be divided into two major sections: (1) The effective classification methods, and (2) the dialog-based object query system (DBOQS) development system.
2.1. The Effective Classification Methods
2.1.1. Chest X-ray and Deep Learning
Karki et al. [
11] and the World Health Organization [
16] proposed a method for classifying DS-TB and DR-TB using a system of deep learning. The CNN architecture ResNet18 was utilized in this research. Radionics [
17] has been utilized to identify patterns that radiologists may have overlooked when examining anomalies in medical images. Multi-task learning [
18] was employed to standardize the core model and to guide the network’s focus on the seemingly relevant portions of the image. To prepare the CXR image for processing, the Original CXR is provided to U-Net, which generates a binary lung mask. The Original CXR is then cropped, and the lungs are segmented inside the bounding box that has been diminished. Tulo et al. [
12] used machine learning (ML) to examine the lungs and mediastinum, in order to classify the TB patients’ drug response. Despite the use of binary categorization [
19] in this study, other kinds of drug-resistant tuberculosis, including DS, MDR, and XDR, are predicted. Several machine learning algorithms, including multi-layer perceptron [
20,
21], K-nearest neighbor [
22,
23], and support vector machine [
23,
24] were used to identify the solution.
Successful deployments of deep learning architectures have been carried out in many fields, from image/video classification to healthcare [
25]. Due in large part to the automatic parsing of radiology reports to generate labels, the CXR research community has profited in recent years from the publication of multiple large, labeled databases, satisfying the insatiable need for data that are inherent to deep learning. In 2017, the NIH clinical center began this pattern by releasing 112,000 photos [
26]. Additional articles later demonstrated that deep learning is superior to other forms of AI in CXR analysis [
27,
28,
29].
2.1.2. Ensemble Deep Learning Model
Ensemble deep learning is a deep learning architecture that uses multiple architectures to address a problem within a single model. Combining the predictions of numerous models has proven to be an excellent way to improve the performances of models. Ensemble learning or ensemble model refers to the process of mixing many predictions from various models, in order to arrive at a final solution. The solution quality of deep learning is contingent on many mechanisms, such as data preparation techniques, deep learning architectures, and fusion techniques.
Image Preprocessing Method
The chest X-ray images used in this research came from seven different countries: Belarus, Georgia, Romania, Azerbaijan, India, Moldova, Ukraine, Kazakhstan, and South Africa. Diverse X-ray machines produce vastly different results in terms of image quality. The initial dimensions of the amassed images varied in the number of pixels. The image analysis workflow highly depends on the pre-processing phase. Through this, it is possible to improve the original image while simultaneously decreasing the amount of noise or distracting elements. Easy image preprocessing methods can dramatically increase the qualities of the classification results of ML and DL. Karki et al. [
11] employed image segmentation to restrict the input pictures for the binary DR/DS classifier to only include regions that are significant for the categorization of pulmonary tuberculosis, i.e., the lungs. In addition, potential confounding elements are eliminated and the lungs are resized in order that they occupy the same percentage of the image. Lung size and patient placement, for example, are commonly connected with clinical settings and could serve as confounding variables. After the segmentation is completed, the results reveal that their proposed model benefits greatly from including these techniques.
Caseneuve et al. [
30] introduced two methods that are referred to as the threshold approach and edge detection to identify CXR images with two clearly discernible lung lobes. These two methods were founded on the concepts of Otsu’s threshold [
31] and the Sobel/Scharr operator [
32,
33]. Sobel examines the presence of the upper/lower image boundaries or left/right image boundaries in the horizontal and vertical gradient calculations. Specifically, the Sobel operator will indicate if visual changes are rapid or gradual. The Scharr operator represents a modification where the mean squared angular error is measured as an extra optimization and an introduction of greater sensitivity. This enables the kernels to be of a derivative type, which approximates Gaussian filters more closely.
Before feeding the images into the CNN model, Ahamed et al. [
34] and Wang et al. [
35] applied the cropping and sharpening filters of image processing techniques to the obtained datasets to enhance the quality of the images, to extract the primary portion of the lung image, i.e., to remove undesirable, irrelevant portions of the photos. The images were trimmed with the correct height-to-width ratio in mind. All of the acquired photos were then filtered with a sharpening filter for enhancement. Conceptually, this filter is derived from Laplacian filters, and is an example of a second-order or second-derivative system of enhancement that accentuates regions of fast intensity variation [
36,
37].
One of the most popular preprocessing techniques that are used to improve the image quality is contrast enhancement. Many contrast enhancement techniques have been introduced to improve the contrast of an image, such as the histogram equalization (HE)-based methods. The classical HE can efficiently utilize displayed intensities, but it tends to over-enhance the contrast if there are high peaks in the histogram, which often results in a harsh and noisy appearance of the output image. Numerous methods have been proposed for limiting the level of enhancement, most of which are obtained through modifications on HE [
35,
38].
In Arici et al. [
39], the authors present a histogram modification framework (HMF) that treats contrast improvement as a minimization of a cost function. To deal with noise and black/white stretching, penalty terms are applied during optimization. Manually adjusting the parameters of HMF allows for varying the degrees of contrast improvement. An automatic image enhancement method based on gamma correction has recently been developed [
40]: Contrast enhancement using adaptive gamma correction with weighting distribution (AGCWD).
If there are high peaks in the input histogram, the AGCWD may be overly simplistic and lead to image blurring in the bright areas. When applied to the entire input image, these global methods are effective, but they overlook the local details. Approaches based on local histogram equalization (LHE) are presented as a means of resolving these shortcomings. The contrast limited adaptive histogram equalization (CLAHE) method proposed by Pizer et al. [
41], is a classic LHE-based image enhancement method that first divides the image into a large number of continuous and non-overlapping sub-blocks, then enhances each sub-block independently, and finally employs an interpolation operation to reduce the block artifacts.
While extreme contrast is possible with the LHE, the final image may be too bright. Furthermore, LHE-based approaches typically necessitate a greater computational effort than GHE-based approaches. Another issue with LHE-based approaches is the dramatic increase in background noise [
41,
42,
43,
44]. The second class of contrast enhancement methods conducts image decomposition before performing enhancement, in order to reduce artifacts and increase subjective quality. Lee et al. [
45] suggested that a gradient domain tone mapping technique can be used. Artifacts may be introduced in non-integrable gradient fields when obtaining the improved image by solving a Poisson equation on the altered gradient field. Edge-aware image processing using the Laplacian pyramid (LPM) was presented by Paris et al. [
46].
CNN Architectures
In addition to the data de-noising and augmentation approaches that may be utilized to improve the classification quality of DL, the convolution neural network (CNN) architecture is one of the most essential characteristics that can considerably increase the classification solution quality. In recent years, numerous CNN architectures have been proposed to improve picture classification performance. CCNet [
47], VGG16, VGG19 [
48], ResNet50 [
49], ResNet101 [
50], DenseNet121 [
51], MobileNetV2 [
28], EfficientNetB1 [
52], NASNetMobile [
53], and EfficientNetB6 [
54] are examples of effective CNN state-of-the-art designs.
Although these types of structures have not yet been applied to the classification of drug-resistant organisms, they have been utilized in a variety of applications. ResNet101 successfully solved the detection of gravelly soil uniformity, and it was superior to all other architectures [
55]. MobileNetV2 is a small, low-latency, low-power model parameterized to satisfy the resource restrictions of a range of use cases. For instance, Zhang et al. [
56] successfully employed MobileNetV2 to determine the fish school feeding behavior. NASNetMobile was proposed by Maharjan [
57] to detect COVID-19 infection in posteroanterior chest X-rays, whereas Chaganti [
58] and Hoorali [
59] used EfficientNetB1 and EfficientNetB6 to classify malware and segment medical images, respectively. Although the abovementioned designs have not yet been used to solve DR-TB and its associated types of DR, it has been demonstrated in numerous publications that they are effective solutions. In this article, a modified version of MobileNetV2, NASNetMobile, EfficientNetB7, and DenseNet121 will be utilized to determine the TB’s drug response.
Decision Fusion Strategies
The effective fusion methodologies utilized to merge the results of several CNN architectures will be employed to improve the solution quality of the TB classification. Decision fusion techniques are the processes that occur when ensemble learning trains many base learners and aggregates their outputs using specified rules. Effective ensemble performance is determined by the rule used to aggregate the outputs. The majority of ensemble models concentrate on the ensemble architecture, and then use naive averaging to forecast the ensemble output. However, the naive averaging of the models, which is followed by the majority of ensemble models, is not data-adaptive, and results in a sub-optimal performance [
24] since it is susceptible to the performances of biased learners.
As there are billions of hyper-parameters in the architecture of deep learning, the issue of overfitting may result in the failures of some base learners. Consequently, techniques, such as the Bayes optimum classifier and super learner, have been adopted to address these problems [
24]. In the literature, the many methods for merging the outputs of ensemble models are as follows: (1) Unweighted model averaging, (2) majority voting, (3) Bayes optimum classifier, (4) stacked generalization, (5) super learner, (6) consensus, and (7) query-by-committee. The unweighted model average weight of all CNNs employed has the same weight, while majority voting uses the decisions of all CNNs to count and select the outcome that the majority of CNNs decide to forecast. Meanwhile, the remaining technique attempts to identify the optimal weight used to make the model’s decision.
Recently, Gonwirat and Surinta [
60] published the differential evolution (DE) approach for automatically optimizing the weighted parameter discovery. The DE algorithm is applied to the weighted parameters, and the optimal weight is then assigned to the ensemble method and stacked ensemble technique. The outcome reveals that the optimal weight finding technique utilizing DE provides a superior answer to the unweighted model averaging and majority voting. Pitakaso et al. [
61] proposed the novel heuristics artificial multiple intelligence system (AMIS) to deal with the network flow problem, and the results suggested that AMIS outperformed DE in locating the superior solution. In this study, we will employ AMIS-ensemble deep learning (AMIS-EDL) to discover the appropriate weight to use as the decision fusion approach rather than the conventional approach.
2.2. DBOQS in Healthcare
Powered by artificial intelligence (AI), DBOQSs are gaining popularity in a variety of businesses, and have substantial application potential in real-world scenarios. However, medical DBOQSs have received little attention, with the majority of published evidence focusing on technology challenges and with limited application research. Moreover, healthcare service providers are interested in adopting new technologies to improve their services [
62], and are beginning to adopt medical DBOQSs for answering/asking questions, creating health records and histories of use, providing information about diseases, discussing the results of clinical tests, and even taking appropriate actions based on users’ responses [
63,
64]. Typically, the DBOQS has been utilized to communicate with patients, in order to maintain the doctor–patient relationship, such as with cancer [
65], mental health [
66,
67], and COVID-19 patients [
68]. It has been demonstrated that the use of DBOQSs decreases the workload of medical personnel and increases patient satisfaction in healthcare [
69].
The DBOQS controller was designed to manage mobile communication between the three CNNs and the users. To detect the health of various crops, including pomegranate trees and firecracker plants, Jain et al. [
70] utilized a CNN with a customized architecture hosted on a cloud service and a mobile application for Android smartphones. Picon et al. [
71] and Esgario et al. [
72] found three wheat illnesses and four coffee leaf diseases and pests, respectively, using a mobile application and a CNN with a modified RetNet-50 architecture hosted on a cloud service. Therefore, the mobile DBOQS is applicable to a variety of real-world issues. Temniranrat et al. [
73] were able to detect five common rice diseases using a CNN with YOLOv3 architecture hosted on a cloud service and LINE, an instant messaging application that is maintained by its development team [
74]. A DBOQS was placed on an instant messaging application (LINE) to access CNNs hosted on a cloud service; therefore, the system offered automatic responses and was always accessible.
From the vast amount of literature, we can conclude that it is possible to use DBOQSs in the field of healthcare image processing, particularly in TB medication response classification, where they only receive an image as input and then report the result using a deep learning model embedded in DBOQS. Using an online DBOQS to inquire about the CXR owner’s chance of developing DR can alleviate the doctors’ workload. Furthermore, this study will provide the recommended treatment plan for the selected patients.
3. Materials and Methods
This study seeks to improve the precision of the deep learning technique used to classify tuberculosis patients’ medication responses. The drug reaction will be divided into four classes, namely DS, DR, MDR, and XDR. After obtaining an effective algorithm to classify them, the DBOQS will be utilized to converse with the doctor, in order to recommend the situation of the patients and the treatment plan for those patients. The therapy schedule recommended by the WHO is indicated in
Figure 1. Consequently, the study approach, depicted in
Figure 2, will be implemented.
The method used to develop the DRCS consists in three steps: (1) Collecting the dataset of the CXRs, as well as the efficacies of previous methods from various literatures; (2) developing the AMIS-EDL algorithm to build the model to classify the drug response; and (3) developing the DRCS using DBOQS. The following is a detailed explanation of each step.
3.1. The Revealed Dataset and Compared Methods
In this experiment, the same dataset was utilized by Karki et al. [
11]. The data collection contains 5019 CXR images associated with tuberculosis, of which 3412 are from DR-TB and 1607 are from DS-TB. It is accessible for download at
https://tbportals.niaid.nih.gov (accessed on 2 August 2022) [
75]. The vast bulk of TB portal data were obtained to identify DR-TB patients from a diverse sample of TB cases. To evaluate the performance, the dataset was partitioned for training and testing at 80% (
n = 4015) and 20% (
n = 1015), respectively.
Table 1 displays the number of DS-TB, DR-TB, MDR-TB, and XDR-TB data used for training and testing.
Table 1 indicates that the total numbers of DS-TB, DR-TB, MDR-TB, and XDR-TB have 1607, 468, 2087, and 857 datasets, respectively; or 32.02%, 9.32%, 41.58%, and 17.08%, respectively. These data will be used to train and test for the accuracy of the proposed methods, compared with the methodology proposed by Ureta and Shrestha [
76], Tulo et al. [
77], Jaeger et al. [
24], Kovalev et al. [
78], Tulo et al. [
12], and Karki et al. [
11]. The details of the compared method are presented in
Table 2.
The key performance indicators (KPIs) to evaluate the suggested methods in comparison to other methods include area under the curve (AUC), F-measure, and accuracy. For binary classification issues, an evaluation metric is the receiver operator characteristic (ROC) curve. The area under the curve (AUC), which serves as a summary of the ROC curve, is a measurement of a classifier’s capacity to distinguish between classes. The harmonic mean of recall and precision is used to calculate the F-measure, giving each the same weight. It enables the evaluation of a model, taking into consideration both precision and recall using a single score, which is useful for explaining the performance of the model and when comparing models. The model’s performance across all classes is often described using its accuracy metric. When every class is equally important, it is useful. It is determined by dividing the total number of guesses by the number of predictions that were correct.
3.2. The Development of Effective Methods
In this study, three methodologies will be utilized to increase the efficacy of the classification of the TB drug response. The data preparation techniques, modern CNN architectures, and decision fusion techniques are these methodologies. Each can be described in greater detail, as follows.
3.2.1. The Data Preprocessing Method
Data preprocessing involves transforming or encoding data in order that they may be easily parsed by a machine. For a model to make accurate and exact predictions, its algorithm must be able to easily interpret the data’s characteristics. Due to their varied origins, the bulk of real-world datasets for machine learning are very prone to be missing, inconsistent, and noisy. The application of data preparation algorithms to these noisy data would not yield excellent results, since the algorithms would be unable to find patterns adequately. Two types of image preprocessing algorithms will be adapted for usage with CXRs, in order to improve the classification accuracy of the proposed model. These two methods are data augmentation and data normalization.
Data Augmentation
By creating synthetic datasets, data augmentation aims to increase the quantity and variety of the training data. The augmented data can be considered to have been derived from a distribution that closely resembles the actual distribution. Then, the expanded dataset can represent features that are more thorough. Image data augmentation techniques can be applied to a variety of data types, including object identification [
79], semantic segmentation [
80], and picture classification [
81].
Image manipulation, image erasure, and image mixing are the fundamental image augmentation techniques. In this research, only image manipulation will be used. Image transformations, such as rotation, mirroring, and cropping, are the focus of fundamental image transformations. The majority of these methods modify images directly, and are simple to implement. Nevertheless, there are downsides. First, it only makes sense to apply fundamental image modifications if the existing data follow a distribution that is near the actual data. Second, some basic image manipulation techniques, such as translation and rotation, suffer from the padding effect. Specifically, following the operation, some sections of the images will be shifted outside the boundary and lost. Consequently, various interpolation methods will be used to fill in the missing data. Typically, the region outside of the image’s border is presumed to be a constant of 0, and will be dark following alteration.
Figure 3 depicts an example of image enhancement employing several types of enhancement, as described earlier.
Image Normalizing Algorithm
Four steps of normalizing the CXR input data have been used. These four steps are shown in
Figure 4.
In
Figure 4, the Original CXRs will be processed with adaptive masking (AM), Gaussian blur (GB), CLAHE, and MVSR. AM is the method depicted in Heidari et al. [
82]. It can execute the procedure by removing the sample’s diaphragm. AM begins by determining the maximum (max) and minimum (min) pixel intensities, followed by the use of threshold techniques for binary thresholding and morphologic closure. This generates the adaptive mask that eliminates the aperture from the source image after a bitwise operation. Gaussian blur (GB) is a filter that operates by calculating a pixel’s value. The filter is based on the normal distribution, which has the form of a bell curve. The concept is that pixels closest to the center pixel carry a greater weight than those further away.
The CLAHE algorithm comprises three main components: Tile generation, histogram equalization, and bilinear interpolation. First, the input image is segmented into pieces. Each unit is referred to as a tile. The input image depicted in the illustration is separated into four tiles. Then, each tile is subjected to histogram equalization using a clip limit that has been specified. Histogram equalization involves five steps: Histogram computation, excess calculation, excess distribution, excess redistribution, and scaling and mapping with a cumulative distribution function (CDF). For each tile, a collection of bins is used to compute the histogram. Histogram bin values exceeding the clip limit are gathered and dispersed over other bins. Then, the CDF for the histogram values is calculated. Using the supplied image pixel values, each tile’s CDF values are scaled and mapped. The generated tiles are stitched together using bilinear interpolation to produce an image with enhanced contrast.
Mean-variance-Softmax-rescale normalization (MVSR normalization) is based on four mathematical operations: The mean of the data, variance, Softmax, and rescaling. According to the probability theory of R. Duncan Luce, sometimes known as Luce’s choice axiom, the probability of one sample being within the same dataset depends on another sample. Typically, the Softmax function is employed in the final activation function of multi-class. In general, the standard deviation (the square root of the variance) is used to create a link between data points and to quantify the spread or distribution of a data collection relative to its mean [
83], using artificial neural network (ANN) models to build an output class probability distribution [
84]. After calculating the normalized intensity of the input, the dataset may include both negative and positive fractional values. The Softmax function was utilized to maintain the impact of negative data and nonlinearity.
To normalize the raw input data into a higher quality of image for use as the input data, in order to increase the performance of the TB classification, image normalization is a crucial component of the image analysis schema. It can improve the original image by minimizing noise or unnecessary elements. In our research, we aligned the training and testing photos by enhancing low-contrast, high-noise input images using four distinct normalization techniques to increase contrast and sharpness.
Figure 5 is an example of when the proposed preprocessing methods are applied.
3.2.2. CNN Architectures
In this research, we will use four compact CNN architectures in an effort to limit the amount of time spent on model construction. As a result, we will use four compact CNN architectures that are nonetheless rather powerful.
MobileNetV2
Sandler et al. [
28] introduced MobileNetV2, which uses depthwise separable convolutional (DwConv) layers and the inverted residuals of the bottleneck block to reduce the weighted parameters of a lightweight network. Depthwise separable convolution is used as a foundational building block in MobileNetV2. However, it adds linear bottlenecks between the layers and shortcut links between the bottlenecks to the design.
Figure 6 depicts the structure of the MobileNetV2’s system.
The bottlenecks encode the model’s intermediate inputs and outputs, while the inner layer embodies the model’s capacity to change from lower-level ideas, such as pixels to higher-level descriptors, such as image categorizations. Finally, as in the case with conventional residual connections, shortcuts allow for faster training and greater precision. MobileNetV2 models are faster over the whole latency spectrum with the same precision. Specifically, the new models use two-fold operations, require 30% fewer parameters, and are around 30% to 40% faster on a Google Pixel phone than MobileNetV1 models, all while attaining more accuracy [
28].
EfficientNetB7
EfficientNet was created by Tan and Le [
52] to search the hyper-parameters of CNN architectures, such as width scaling, depth scaling, resolution scaling, and compound scaling. In addition, squeeze-and-excitation (SE) optimization was added to the bottleneck block of EfficientNet, in order to generate an informative channel feature with GAP summation. Then, correlation features are identified by lowering the dimensions to small sizes and changing them back to their original size. EfficientNetB1 was developed based on MobileNetV2; however, its resolutions, channels, and repetition rates vary. EfficientNetB1 is comparable to MobileNetV2, which replaced Conv1 with 512 feature maps and eliminated the fourth bottleneck block.
Figure 7 depicts the architecture of the EfficientNetB1 overall architecture, which may be separated into seven blocks. Each block of MBConv’s associated filter size is displayed.
DenseNet121
In a conventional feed-forward convolutional neural network (CNN), each convolutional layer, with the exception of the first, obtains the output of the previous convolutional layer and generates an output feature map that is then transferred to the next convolutional layer. Therefore, for “L” layers, there are “L” direct connections, one between each layer and the following one. However, when the number of layers in the CNN increases, i.e., as they become deeper, the “vanishing gradient” issue develops. This implies that as the channel for information from the input to output layers becomes longer, certain information may “disappear” or become lost; therefore, reducing the network’s capacity to train successfully. DenseNets solve this issue by changing the traditional CNN design and streamlining the interlayer connectivity structure; therefore, the moniker “densely connected convolutional network”. L(L + 1)/2 direct connections exist between “L” levels. In each layer, the feature maps of all preceding layers are not added together, but rather, they are concatenated and used as inputs. As a result, DenseNets require fewer parameters than a comparable standard CNN, which enables feature reuse as redundant feature mappings are removed.
When the sizes of feature maps vary, it is not possible to use the concatenation method. Nevertheless, a crucial component of CNNs is the down sampling of layers, which minimizes the size of feature maps via dimensionality reduction, in order to increase computation speeds. To achieve this, DenseNets are subdivided into DenseBlocks, where the dimensions of the feature maps remain constant inside a block, but the number of filters between blocks varies. Transition layers are the layers between the blocks that lower the number of channels by half. For each layer, the above equation is solved.
An example of a deep DenseNet with three dense blocks is displayed in
Figure 8; within the dense block, the feature maps are all the same size, in order that the features can be concatenated, while the transition layers between neighboring blocks conduct down sampling via convolution and pooling procedures. DenseNet121 comprises 7 × 7 convolutions and 3 × 3 max pooling. It has
dense blocks (3) and
dense blocks, (4) while for other dense blocks, it has the same as DenseNet169, DenseNet201, and DenseNet264.
NASNetMobile
The Google brain team created the neural architecture search network (NASNet), which has two key functions: Normal cell and reduction cell. In order to achieve a higher map, NASNet initially applies its operations to the small dataset before transferring its block to the large dataset. For optimal regularization, the speed of NASNet is increased via a customized drop path known as the scheduled drop path. In the original NASNet architecture, where the number of cells is not predetermined, normal and reduction cells are used [
85], with the normal cells determining the feature map size and the reduction in the cells returning the feature map with its height and width decreased by a factor of two. A recurrent neural network (RNN)-based control architecture in NASNet predicts the entire network topology using two initial hidden states. The controller architecture uses an RNN-based LSTM model with Softmax prediction for convolutional cell prediction and recursively produced network motifs. NASNetMobile accepts images with a resolution of 224 by 224 pixels, whereas NASNetLarge accepts images with a resolution of 331 by 331 pixels. NASNetMobile leverages the pre-trained ImageNet network weights for transfer learning, in order to recognize the objects.
Figure 9 shows the architecture of NASNetMobile.
All the architectures that are previously explained will be used in both heterogeneous and homogeneous manners in the ensemble deep learning, as shown in
Figure 10.
3.2.3. Decision Fusion Strategy
The decision fusion strategy (DFS) combines the classification decisions of many classifiers into a single conclusion about the event. When it is challenging to combine all of the CNN’s results, several classifiers are frequently employed with multi-modal CNN. Several viable DFSs for application in deep learning ensembles, including the unweighted model average, majority voting, and various weight optimization techniques, were discussed in [
86]. The designed methodologies for this investigation will employ three approaches. These methods are the unweighted model average and majority voting, as described in Johnson et al. [
87], together with our proposed DFS, which is the AMIS-DFS developed by Pitakaso et al. [
61].
Unweighted Model Average
The most prevalent method for fusing decisions in the literature is the unweighted average of the outputs of the base learners in an ensemble. The final prediction of the ensemble model is obtained by averaging the results of the base learners. Since deep learning architectures have a high variance and a low bias, a simple averaging of the ensemble models improves the generalization performance by reducing the variance among the models. The averaging of the base learners is conducted directly on the outputs of the base learners, or on the predicted probabilities of the classes using the Softmax function [
88] illustrated in Equation (1).
where
is the probability of the
i-th unit on the
j-th base learner,
is the output of the ith unit of the
j-th base learner, and
K is the number of classes.
Majority Voting
Similar to unweighted averaging, majority voting aggregates the base learners’ outputs. In contrast, majority voting counts the votes of the base learners and predicts the final labels as the label that received the majority of votes. In contrast to unweighted averaging, majority voting is less skewed towards the outcome of a certain learner, since the influence is offset by a majority vote count. However, if the majority of similar or dependent base learners favors a certain event, that event will predominate in the ensemble model. In majority voting, Kuncheva et al. [
89] found that pairwise dependence among the base learners plays a significant influence, whereas for the categorization of images, the prediction of shallow networks is more diversified than those of deeper networks [
90].
Optimization of Weight Using AMIS-DFS
The principle underlying AMIS is to use a system with many artificial intelligences to aid in discovering optimal solutions. “Intelligence box or IB” refers to a collection of computing methods or algorithms with unique properties. AMIS is a heuristic based on population. A population member is known as a work package (WP). Each WP will independently enhance its solution using the chosen IB. AMIS will generate a baseline set of work packages (WP), WP will select IB to improve its own solution in each iteration, update its heuristics’ information, and repeat this cycle until the termination condition is met. The AMIS can be elaborated upon as follows.
Initial Work Package Generation
At random, initial work packages (WP) will be generated. In this phase, the NP number of WPs is displayed. The number of CNN defines dimension D of the WP. The first set of WP is illustrated in
Table 3. Let
Xijt stand for the first set of
WPi (
i = 1 in this case), element
j iteration
t.
At iteration
t, the weight that will be used to classify the drug response for a CXR can be determined using Equation (2).
Figure 11 depicts the framework for calculating the total weight using the value of element
j of
WPi at iteration
t.
Let
be the average value of the prior classification value for a specific
i and
t.
can be determined with Equation (2).
When equals as determined via the AMIS procedure, i is the number of WP, j is the number of CNN or the number of WP elements, and t is the iteration counter.
3.3. DBOQS Design
The DBOQS was developed to provide the medical team with an answer to the question of whether or not the patient has the probability of receiving the DR.
Figure 13 depicts the framework of the DBOQS that we designed for this project.
From
Figure 13, when a doctor (user) wants to use the DBOQS to inquire about the drug response of the TB patients, the doctor must enable the LINE Application on their mobile phone to scan the QR code to add DBOQS as a friend for the first time. Then, the doctors can register and use the DBOQS, which has a basic menu for the doctor to use. Once the doctor has selected an item from the menu (Message input), the LINE Application will send a message request to the LINE platform that provides the DBOQS service for checking the message event received from the LINE Application. Once completed, they will send a message reply to the LINE platform, and the LINE platform will send the message back to the LINE Application for the doctor to read the message. For the administrator who developed the model, the model obtained from
Section 3.2 will be the model that will be used in the DBOQS. Moreover, information with regard to the treatment program will be inputted as the information given to the user.
The DBOQS or the LINE Messaging API was collaboratively applied with CNN through the Flask API for assisting with the diagnosis of DR-TB. Once DBOQS receives an X-ray image that is sent by the user, the DBOQS will transmit the image to the DBOQS server’s webhook, and then the image is sent to the deep learning model. The deep learning models will predict the disease in the submitted image. Drug recommendations for predictable diseases are then extracted, and predictable disease outcomes and explanations are sent in text via the LINE Messaging API and LINE platform. This system can work for 24 h per day in real-time.
4. Computational Results
In this section, the computational results will be divided into two main parts, which are: (1) The test to reveal the effectiveness of the proposed methods, and (2) the result of developing DBOQS for the drug response classification.
4.1. Revealing the Effectiveness of the Proposed Models
This section will be divided into two sub-sections, which are: (1) Comparing the efficiency among all proposed models, in order to have multi-class classification and (2) comparing the classification of the proposed model with an existing method that has binary classification. All comparisons will use the same dataset, as explained in
Section 3.1.
4.1.1. Comparing the Effectiveness of the Proposed Models
All 60 methods will be evaluated in multi-class classification quality using the classification of four classes, and the results of all 60 methods are shown in
Table 6.
From the results obtained in
Table 6, we can see that M-12 obtained the best results of all key performance measurements. The confusion matrices of the classification are shown in
Figure 14. The confusion matrices reveal that when employing a multi-class classification model to categorize DS-TB, the majority of incorrect results fall into the MDR-TB class, whereas the other two classes have the same percentage of incorrect results. There is no major difference between the remaining classification ambiguity. Using the binary classification approach, DS-TB has the highest proportion of incorrect predictions when compared pairwise with other classes. However, the classification of MDR-TB and XDR-TB is the most perplexing, with 15.43% of incorrect classifications.
From
Table 7, using no preprocessing technique yields a worse solution than using all the preprocessing techniques (10.05%). Using all the CNN architectures in the ensemble yields a result that is 6.42% better than using the same CNN architecture in the ensemble. Finally, using AMIS-EDL yields a result that is 1.17% better than using UMA and MV as the decision fusion strategy. Utilizing all data augmentation approaches and data normalization in conjunction with all CNN architectures (heterogeneous) and AMIS-EDL as the decision fusion mechanism will be our plan for the next experiment.
4.1.2. Comparing the Effectiveness of the Proposed Model with Existing Methods
In this experiment, we perform a comparison of the best method found in
Section 4.1 with the existing heuristics. The comparison will be performed as the binary classification, and the result is shown in
Table 8,
Table 9,
Table 10 and
Table 11.
Comparing the classification of DS-TB and DR-TB based on the results in
Table 8, the AUCs of the proposed model are 43.13% and 21.39% better than those of Ureta and Shrestha [
75], and of Karki et al. [
11]. Using F-measure as the criterion, the proposed model is 1.09% better than the model of Tulo et al. [
76]. When classifying between DS-TB and DR-TB, we can conclude that the suggested method is 30.01% more accurate than the alternative methods. Comparing the classification results of DR-TB and MDR-TB between the proposed approaches of Jaeger et al. [
24] and Tulo et al. [
12], utilizing all KPIs, the proposed method provides a superior answer by 2.52–46.27% (using the information provided in
Table 9). The classification results of DR-TB and XDR-TB utilizing the data presented in
Table 10 demonstrated that the proposed approach is 1.71–5.52% superior to the method of Tulo et al. [
12]. Finally,
Table 11 reveals that the proposed approaches produce solutions that are 2.43% to 6.67% more accurate than the method of Tulo et al.
4.2. The DBOQS Result
DBOQS has been developed using the concept explained in
Section 3.3, and the model developed in
Section 4.1, which is model AD-10, will be used as the model for DBOQS. The treatment program suggested is the WHO recommendation program, as we mentioned in
Section 1. An example of the Q and A results of DBOQS is shown in
Figure 15.
We perform the experiment in DBOQS, and input the CXR 443 images (randomly selected from the test and train dataset) to identify its accuracy and other performance measures; the results are shown in
Table 12.
The results in
Table 12 indicate that the proper categorization rate is 90.74%, which is not significantly different from the results in
Section 4.1. DS, DR, MDR, and XDR had respective accurate answers of 90.38%, 90.08%, 91.30%, and 91.26%. It is demonstrated that each class has a classification accuracy of at least 90%, a level that is sufficient for clinicians to place their trust in the classifications.
In the next experiment, we let 30 doctors try our DBOQS and evaluate them via a questionnaire; the results of the questionnaire are presented in
Table 13. The scores from the doctors were averaged. A score equal to 1 indicated that the doctors did not agree with the question; when the score was 10, this indicated that the doctor strongly agreed with the question.
Table 13 shows that users rated the DBOQS 9.51 out of 10 for its response time, suggesting that they found it to be extremely helpful. With a mean score of 9.58, people agreed that the DBOQS provides the correct answer when asked about drug resistance. With a score of 9.97 and 9.86, respectively, they believed that the DRCS would make it simpler to diagnose the drug reaction and lessen their burden. After the testing phase of the study, participants rated their trust in and willingness to utilize the DRCS as high as 9.51 and 9.97, respectively.
5. Discussion
The developed approach in this study is one of the first to classify TB patients into four distinct groups based on their responses to the treatment. DS-TB, DR-TB, MDR, and XDR are the subtypes that are currently known. Since there are many possible ways to classify a patient’s reaction to a medicine, a multi-class classification model was used. Two additional studies, Karki et al. [
11] and Tulo et al. [
12], also divide TB patients into four groups, but they achieve this using a binary classification system (BC). Every time BC is used, it will always provide a 0 or a 1 for the outcome, such as when comparing DS and DR. Furthermore, our algorithms can be classified into four classes simultaneously, making this classification model more user-friendly. Once we know the type of TB, we can more quickly plan the patient’s treatment schedule.
Additionally, the AMIS-EDL model was created to enhance the quality of the solutions provided using pre-existing models. Computational results demonstrate that AMIS-EDL outperforms the current approaches, including the ones proposed by Karki et al. [
11] and Tulo et al. [
12]. This is due to the fact that the AMIS algorithm, as developed by Pitakaso et al. [
61], has been shown to be effective. For example, it can obtain better results than the genetic algorithm and the differential evolution algorithm, two of the most popular current heuristics. The suggested model outperforms the conventional DL method, which relies on the unweighted model average (UMA) and majority voting (MV) by a profit of 31.25%. This follows the reasoning of Gonwirat and Surinta [
91], who similarly concluded that metaheuristics, such as AMIS, can enhance the quality of solutions obtained using DLs employing the conventional UMA and MV.
The user has confidence in the DRCS’s ability to generate a credible result of 95.1%, according to the computational result. As a result of the DBOQS’s role, the medical staff usually do not give it a sufficient amount of chance for employment in medical care due to low trust and the appearance of unreliability. Medical staff were overstretched due to the increasing need for hospital care during the COVID-19 pandemic. The accuracy and dependability of a DBOQS are key factors in increasing the likelihood that medical professionals will choose to utilize it as an aid. This is why our system has a good chance of being adopted and utilized by medical experts to aid in the diagnosis and treatment planning of tuberculosis.
As we know from the results, the DRCS has more than 90% accuracy, and it can improve precision from traditional and existing methods by 1.17% to 43.43%. Therefore, when the accuracy is high and the workload is high, it would be good for the medical staff to use the DRCS for reducing their workload, and it would increase the service quality of the medical staff, as well.
6. Conclusions and Outlook
This article develops effective methods for classifying DS-TB, DR-TB, MDR, and XDR patients. Using ensemble deep learning, the desired outcome has been predicted. We selected four small and effective CNN architectures for deep learning; namely, MobileNetV2, EfficientNetB7, NASNetMobile, and DenseNet121. In addition, image augmentation and a normalization technique were incorporated into the model to improve the quality of the proposed methods. Finally, unweighted ensemble averaging, majority voting, and AMIS-EDL were used to discover the optimal weight parameters for the ensemble deep learning system. The proposed methods were compared to the performances of several types of methods addressed in previous studies.
The computational findings demonstrate that, among the deep learning architectures offered, EfficientNetB7 offers the most effective solution. It improves the solution by 1.76–2.41% compared to the others. The best architecture for classifying the DR is the one that combines all the architectures into a single model. The optimum architecture for classifying DR utilizes all the CNN architectures inside a single classification model. Using multiple CNN architectures can enhance the quality (accuracy) by 5.21–7.75%. The preprocessing techniques increase the quality of the answer by an average of 10.41% when compared to methods that do not use preprocessing. In comparison to other fusion methods, AMIS can improve the solution quality by 1.17%.
In comparison to the method described by Karki et al. [
11], the solution quality can be enhanced by 31.25%. Compared to the other approaches proposed in a previous study, the proposed method can provide solutions that are 2.87–7.15% better than Tulo et al. [
12], 48.58% better than Jaeger et al. [
24], 53.16% better than Kovalev et al. [
77], and 43.13% better than Ureta and Shrestha [
75]. We can conclude, based on this result, that the proposed method surpassed all of the other methods proposed in the literature in terms of producing more exact classifications. The computational results demonstrate that the DBOQS developed for use in the DRCS can predict drug responses with an accuracy of more than 90% in every group of drug response, and the overall accuracy is 90.74%. Overall, with a result of 9.51 points out of 10 points, the clinicians who used the developed system had faith in it, and 99.7% of the doctors planned to keep using it as a secondary diagnostic tool in the future.
It is necessary to further develop the DRCS in order that it can: (1) Classify the various forms of drug resistance; (2) recommend a treatment regimen; (3) monitor the patient’s progress after beginning the recommended regimen; and (4) collect data from the user in order to analyze and recommend a subsequent treatment plan. Despite this, data mining and two-way communication DBOQSs are vital for the DRCS to make informed decisions and to provide meaningful advice to medical professionals when necessary.
Author Contributions
Conceptualization, C.P., R.P., T.P. and N.N.; methodology, R.P. and S.G.; software, S.G., P.E. and C.K.; validation, S.S.J. and N.W.; formal analysis, S.G., T.S. and S.K.; investigation, P.E.; resources, T.S.; data curation, S.S.J., N.W. and S.G.; writing—original draft preparation, C.P. and R.P.; writing—review and editing, T.P. and N.N.; visualization, C.K. and S.K.; supervision, C.P.; project administration, C.P.; funding acquisition, T.S. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
This research was supported by the Ministry of Higher Education, Science, Research, and Innovation (MHESRI), and the Artificial Intelligence Optimization SMART laboratory at Ubon Ratchathani University.
Conflicts of Interest
The authors declare no conflict of interest.
References
- World Health Organization. Global Tuberculosis Report 2020; WHO: Geneva, Switzerland, 2020.
- Mignani, S.; Tripathi, R.P.; Chen, L.; Caminade, A.-M.; Shi, X.; Majoral, J.-P. New ways to treat tuberculosis using dendrimers as nanocarriers. Pharmaceutics 2018, 10, 105. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- World Health Organization. WHO Consolidated Guidelines on Drug-Resistant Tuberculosis Treatment; WHO: Geneva, Switzerland, 2019.
- World Health Organization. WHO Operational Handbook on Tuberculosis: Module 4: Treatment: Drug-Susceptible Tuberculosis Treatment; WHO: Geneva, Switzerland, 2022.
- Laya, B.; Domingo, M.; Javier, X.; Sanches, M. Drug resistant tuberculosis radiologic imaging manifestation. TB corner. Philippines 2015, 1, 1–5. [Google Scholar]
- World Health Organization. Chest Radiography in Tuberculosis Detection: Summary of Current WHO Recommendations and Guidance on Programmatic Approaches; 9241511508; WHO: Geneva, Switzerland, 2016.
- Gomes, M.; Saad, R., Jr.; Stirbulov, R. Pulmonary tuberculosis: Relationship between sputum bacilloscopy and radiological lesions. Rev. Do Inst. Med. Trop. São Paulo 2003, 45, 275–281. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Jamzad, A.; Khatami, A. Radiographic findings of pulmonary tuberculosis in Tehran in comparison with other institutional studies. Iran. J. Radiol. 2009, 6, e78964. [Google Scholar]
- Qin, Z.Z.; Sander, M.S.; Rai, B.; Titahong, C.N.; Sudrungrot, S.; Laah, S.N.; Adhikari, L.M.; Carter, E.J.; Puri, L.; Codlin, A.J. Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems. Sci. Rep. 2019, 9, 15000. [Google Scholar] [CrossRef] [Green Version]
- Qin, Z.Z.; Ahmed, S.; Sarker, M.S.; Paul, K.; Adel, A.S.S.; Naheyan, T.; Barrett, R.; Banu, S.; Creswell, J. Tuberculosis detection from chest X-rays for triaging in a high tuberculosis-burden setting: An evaluation of five artificial intelligence algorithms. Lancet Digit. Health 2021, 3, e543–e554. [Google Scholar] [CrossRef] [PubMed]
- Karki, M.; Kantipudi, K.; Yang, F.; Yu, H.; Wang, Y.X.J.; Yaniv, Z.; Jaeger, S. Generalization Challenges in Drug-Resistant Tuberculosis Detection from Chest X-rays. Diagnostics 2022, 12, 188. [Google Scholar] [CrossRef]
- Tulo, S.; Ramu, P.; Swaminathan, R. Evaluation of Diagnostic Value of Mediastinum for Differentiation of Drug Sensitive, Multi and Extensively Drug Resistant Tuberculosis Using Chest X-rays. IRBM 2022, in press. [Google Scholar] [CrossRef]
- Han, J.; Kamber, M. Data Mining: Concepts and Techniques; The Morgan Kaufinann Series in Data Management Systems; Jim, G., Ed.; Morgan Kaufmann Publishers: Burlington, MA, USA, 2000; Volume 10, p. 2020. [Google Scholar]
- Wu, Y.; Ianakiev, K.; Govindaraju, V. Improved k-nearest neighbor classification. Pattern Recognit. 2002, 35, 2311–2318. [Google Scholar] [CrossRef]
- Cunningham, P.; Delany, S.J. K-nearest neighbour classifiers-a tutorial. ACM Comput. Surv. 2021, 54, 1–25. [Google Scholar] [CrossRef]
- World Health Organization. Global Tuberculosis Report 2021; WHO: Geneva, Switzerland, 2021.
- Van Griethuysen, J.J.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.; Fillion-Robin, J.-C.; Pieper, S.; Aerts, H.J. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017, 77, e104–e107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar]
- Toğaçar, M.; Ergen, B.; Cömert, Z.; Özyurt, F. A deep feature learning model for pneumonia detection applying a combination of mRMR feature selection and machine learning models. IRBM 2020, 41, 212–222. [Google Scholar] [CrossRef]
- Govindarajan, S.; Swaminathan, R. Analysis of tuberculosis in chest radiographs for computerized diagnosis using bag of keypoint features. J. Med. Syst. 2019, 43, 1–9. [Google Scholar] [CrossRef]
- Krishna, K.D.; Akkala, V.; Bharath, R.; Rajalakshmi, P.; Mohammed, A.; Merchant, S.; Desai, U. Computer aided abnormality detection for kidney on FPGA based IoT enabled portable ultrasound imaging system. IRBM 2016, 37, 189–197. [Google Scholar] [CrossRef]
- Kukker, A.; Sharma, R. Modified fuzzy Q learning based classifier for Pneumonia and tuberculosis. IRBM 2021, 42, 369–377. [Google Scholar] [CrossRef]
- Ramaniharan, A.K.; Manoharan, S.C.; Swaminathan, R. Laplace Beltrami eigen value based classification of normal and Alzheimer MR images using parametric and non-parametric classifiers. Expert Syst. Appl. 2016, 59, 208–216. [Google Scholar] [CrossRef]
- Jaeger, S.; Juarez-Espinosa, O.H.; Candemir, S.; Poostchi, M.; Yang, F.; Kim, L.; Ding, M.; Folio, L.R.; Antani, S.; Gabrielian, A. Detecting drug-resistant tuberculosis in chest radiographs. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 1915–1925. [Google Scholar] [CrossRef] [Green Version]
- Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
- Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
- Qin, C.; Yao, D.; Shi, Y.; Song, Z. Computer-aided detection in chest radiography based on artificial intelligence: A survey. Biomed. Eng. Online 2018, 17, 1–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
- Anis, S.; Lai, K.W.; Chuah, J.H.; Ali, S.M.; Mohafez, H.; Hadizadeh, M.; Yan, D.; Ong, Z.-C. An overview of deep learning approaches in chest radiograph. IEEE Access 2020, 8, 182347–182354. [Google Scholar] [CrossRef]
- Caseneuve, G.; Valova, I.; LeBlanc, N.; Thibodeau, M. Chest X-Ray Image Preprocessing for Disease Classification. Procedia Comput. Sci. 2021, 192, 658–665. [Google Scholar] [CrossRef]
- Zhang, J.; Hu, J. Image segmentation based on 2D Otsu method with histogram analysis. In Proceedings of the 2008 International Conference on Computer Science and Software Engineering, Washington, DC, USA, 12–14 December 2008; pp. 105–108. [Google Scholar]
- Farid, H.; Simoncelli, E.P. Optimally rotation-equivariant directional derivative kernels. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Kiel, Germany, 10–12 September 1997; pp. 207–214. [Google Scholar]
- Scharr, H. Optimal filters for extended optical flow. In Proceedings of the International Workshop on Complex Motion, Günzburg, Germany, 12–14 October 2004; pp. 14–29. [Google Scholar]
- Ahamed, K.U.; Islam, M.; Uddin, A.; Akhter, A.; Paul, B.K.; Yousuf, M.A.; Uddin, S.; Quinn, J.M.; Moni, M.A. A deep learning approach using effective preprocessing techniques to detect COVID-19 from chest CT-scan and X-ray images. Comput. Biol. Med. 2021, 139, 105014. [Google Scholar] [CrossRef] [PubMed]
- Wang, C.; Peng, J.; Ye, Z. Flattest histogram specification with accurate brightness preservation. IET Image Process. 2008, 2, 249–262. [Google Scholar] [CrossRef]
- Bhairannawar, S.S. Efficient medical image enhancement technique using transform HSV space and adaptive histogram equalization. In Soft Computing Based Medical Image Analysis; Elsevier: Amsterdam, The Netherlands, 2018; pp. 51–60. [Google Scholar]
- Woods, J.W. Multidimensional Signal, Image, and Video Processing and Coding, 2nd ed.; Academic Press: Cambridge, MA, USA, 2011. [Google Scholar]
- Lee, C.; Lee, C.; Lee, Y.-Y.; Kim, C.-S. Power-constrained contrast enhancement for emissive displays based on histogram equalization. IEEE Trans. Image Process. 2011, 21, 80–93. [Google Scholar] [PubMed]
- Arici, T.; Dikbas, S.; Altunbasak, Y. A histogram modification framework and its application for image contrast enhancement. IEEE Trans. Image Process. 2009, 18, 1921–1935. [Google Scholar] [CrossRef] [PubMed]
- Huang, S.-C.; Cheng, F.-C.; Chiu, Y.-S. Efficient contrast enhancement using adaptive gamma correction with weighting distribution. IEEE Trans. Image Process. 2012, 22, 1032–1041. [Google Scholar] [CrossRef] [PubMed]
- Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
- Dale-Jones, R.; Tjahjadi, T. A study and modification of the local histogram equalization algorithm. Pattern Recognit. 1993, 26, 1373–1381. [Google Scholar] [CrossRef]
- Stark, J.A. Adaptive image contrast enhancement using generalizations of histogram equalization. IEEE Trans. Image Process. 2000, 9, 889–896. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Kim, T.K.; Paik, J.K.; Kang, B.S. Contrast enhancement system using spatially adaptive histogram equalization with temporal filtering. IEEE Trans. Consum. Electron. 1998, 44, 82–87. [Google Scholar]
- Lee, W.-F.; Lin, T.-Y.; Chu, M.-L.; Huang, T.-H.; Chen, H.H. Perception-based high dynamic range compression in gradient domain. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1805–1808. [Google Scholar]
- Paris, S.; Hasinoff, S.W.; Kautz, J. Local laplacian filters: Edge-aware image processing with a laplacian pyramid. ACM Trans. Graph. 2011, 30, 68. [Google Scholar] [CrossRef]
- Chamchong, R.; Gao, W.; McDonnell, M.D. Thai handwritten recognition on text block-based from Thai archive manuscripts. In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, Australia, 20–25 September 2019; pp. 1346–1351. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Shinohara, I.; Inui, A.; Mifune, Y.; Nishimoto, H.; Mukohara, S.; Yoshikawa, T.; Kuroda, R. Ultrasound with Artificial Intelligence Models Predicted Palmer 1B Triangular Fibrocartilage Complex Injuries. Arthrosc. J. Arthrosc. Relat. Surg. 2022, 38, 2417–2424. [Google Scholar] [CrossRef] [PubMed]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
- Khan, S.I.; Shahrior, A.; Karim, R.; Hasan, M.; Rahman, A. MultiNet: A deep neural network approach for detecting breast cancer through multi-scale feature fusion. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6217–6228. [Google Scholar] [CrossRef]
- Li, R.; St George, R.J.; Wang, X.; Lawler, K.; Hill, E.; Garg, S.; Williams, S.; Relton, S.; Hogg, D.; Bai, Q. Moving towards intelligent telemedicine: Computer vision measurement of human movement. Comput. Biol. Med. 2022, 147, 105776. [Google Scholar] [CrossRef] [PubMed]
- Qu, X.; Wang, J.; Wang, X.; Hu, Y.; Zeng, T.; Tan, T. Gravelly soil uniformity identification based on the optimized Mask R-CNN model. Expert Syst. Appl. 2022, 212, 118837. [Google Scholar] [CrossRef]
- Zhang, L.; Wang, J.; Li, B.; Liu, Y.; Zhang, H.; Duan, Q. A MobileNetV2-SENet-based method for identifying fish school feeding behavior. Aquac. Eng. 2022, 99, 102288. [Google Scholar] [CrossRef]
- Maharjan, J.; Calvert, J.; Pellegrini, E.; Green-Saxena, A.; Hoffman, J.; McCoy, A.; Mao, Q.; Das, R. Application of deep learning to identify COVID-19 infection in posteroanterior chest X-rays. Clin. Imaging 2021, 80, 268–273. [Google Scholar] [CrossRef]
- Chaganti, R.; Ravi, V.; Pham, T.D. Image-based malware representation approach with EfficientNet convolutional neural networks for effective malware classification. J. Inf. Secur. Appl. 2022, 69, 103306. [Google Scholar] [CrossRef]
- Hoorali, F.; Khosravi, H.; Moradi, B. IRUNet for medical image segmentation. Expert Syst. Appl. 2022, 191, 116399. [Google Scholar] [CrossRef]
- Gonwirat, S.; Surinta, O. Optimal weighted parameters of ensemble convolutional neural networks based on a differential evolution algorithm for enhancing pornographic image classification. Eng. Appl. Sci. Res. 2021, 48, 560–569. [Google Scholar]
- Pitakaso, R.; Nanthasamroeng, N.; Srichok, T.; Khonjun, S.; Weerayuth, N.; Kotmongkol, T.; Pornprasert, P.; Pranet, K. A Novel Artificial Multiple Intelligence System (AMIS) for Agricultural Product Transborder Logistics Network Design in the Greater Mekong Subregion (GMS). Computation 2022, 10, 126. [Google Scholar] [CrossRef]
- Montenegro, J.L.Z.; da Costa, C.A.; da Rosa Righi, R. Survey of conversational agents in health. Expert Syst. Appl. 2019, 129, 56–67. [Google Scholar] [CrossRef]
- Almalki, M.; Azeez, F. Health chatbots for fighting COVID-19: A scoping review. Acta Inform. Med. 2020, 28, 241. [Google Scholar] [CrossRef]
- van Heerden, A.; Ntinga, X.; Vilakazi, K. The potential of conversational agents to provide a rapid HIV counseling and testing services. In Proceedings of the 2017 International Conference on the Frontiers and Advances in Data Science (FADS), Xi’an, China, 23–25 October 2017; pp. 80–85. [Google Scholar]
- Chaix, B.; Bibault, J.-E.; Pienkowski, A.; Delamon, G.; Guillemassé, A.; Nectoux, P.; Brouard, B. When chatbots meet patients: One-year prospective study of conversations between patients with breast cancer and a chatbot. JMIR Cancer 2019, 5, e12856. [Google Scholar] [CrossRef] [PubMed]
- De Gennaro, M.; Krumhuber, E.G.; Lucas, G. Effectiveness of an empathic chatbot in combating adverse effects of social exclusion on mood. Front. Psychol. 2020, 10, 3061. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Fulmer, R.; Joerin, A.; Gentile, B.; Lakerink, L.; Rauws, M. Using psychological artificial intelligence (Tess) to relieve symptoms of depression and anxiety: Randomized controlled trial. JMIR Ment. Health 2018, 5, e9782. [Google Scholar] [CrossRef] [PubMed]
- Martin, A.; Nateqi, J.; Gruarin, S.; Munsch, N.; Abdarahmane, I.; Zobel, M.; Knapp, B. An artificial intelligence-based first-line defence against COVID-19: Digitally screening citizens for risks via a chatbot. Sci. Rep. 2020, 10, 19012. [Google Scholar] [CrossRef]
- Xu, L.; Sanders, L.; Li, K.; Chow, J.C. Chatbot for health care and oncology applications using artificial intelligence and machine learning: Systematic review. JMIR Cancer 2021, 7, e27850. [Google Scholar] [CrossRef]
- Jain, L.; Vardhan, H.; Nishanth, M.; Shylaja, S. Cloud-based system for supervised classification of plant diseases using convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Cloud Computing in Emerging Markets (CCEM), Bangalore, India, 1–3 November 2017; pp. 63–68. [Google Scholar]
- Picon, A.; Alvarez-Gila, A.; Seitz, M.; Ortiz-Barredo, A.; Echazarra, J.; Johannes, A. Deep convolutional neural networks for mobile capture device-based crop disease classification in the wild. Comput. Electron. Agric. 2019, 161, 280–290. [Google Scholar] [CrossRef]
- Esgario, J.G.; de Castro, P.B.; Tassis, L.M.; Krohling, R.A. An app to assist farmers in the identification of diseases and pests of coffee leaves using deep learning. Inf. Process. Agric. 2022, 9, 38–47. [Google Scholar] [CrossRef]
- Temniranrat, P.; Kiratiratanapruk, K.; Kitvimonrat, A.; Sinthupinyo, W.; Patarapuwadol, S. A system for automatic rice disease detection from rice paddy images serviced via a Chatbot. Comput. Electron. Agric. 2021, 185, 106156. [Google Scholar] [CrossRef]
- Cheng, H.-H.; Dai, Y.-L.; Lin, Y.; Hsu, H.-C.; Lin, C.-P.; Huang, J.-H.; Chen, S.-F.; Kuo, Y.-F. Identifying tomato leaf diseases under real field conditions using convolutional neural networks and a chatbot. Comput. Electron. Agric. 2022, 202, 107365. [Google Scholar] [CrossRef]
- Rosenthal, A.; Gabrielian, A.; Engle, E.; Hurt, D.E.; Alexandru, S.; Crudu, V.; Sergueev, E.; Kirichenko, V.; Lapitskii, V.; Snezhko, E.; et al. The TB Portals: An Open-Access, Web-Based Platform for Global Drug-Resistant-Tuberculosis Data Sharing and Analysis. J. Clin. Microbiol. 2017, 55, 3267–3282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ureta, J.; Shrestha, A. Identifying drug-resistant tuberculosis from chest X-ray images using a simple convolutional neural network. J. Phys. Conf. Ser. 2021, 2071, 012001. [Google Scholar] [CrossRef]
- Tulo, S.K.; Ramu, P.; Swaminathan, R. An Automated Approach to Differentiate Drug Resistant Tuberculosis in Chest X-ray Images Using Projection Profiling and Mediastinal Features. In Public Health and Informatics; IOS Press: Amsterdam, The Netherlands, 2021; pp. 512–513. [Google Scholar]
- Kovalev, V.; Liauchuk, V.; Kalinovsky, A.; Rosenthal, A.; Gabrielian, A.; Skrahina, A.; Astrauko, A.; Tarasau, A.; Kalinouski, A.; Rosenthal, A. Utilizing radiological images for predicting drug resistance of lung tuberculosis. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 1–312. [Google Scholar]
- Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X. Deep learning for generic object detection. A Survey. arXiv 2018, arXiv:1809.02165. [Google Scholar] [CrossRef] [Green Version]
- Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3544. [Google Scholar] [CrossRef] [PubMed]
- Algan, G.; Ulusoy, I. Image classification with deep learning in the presence of noisy labels: A survey. Knowl. Based Syst. 2021, 215, 106771. [Google Scholar] [CrossRef]
- Heidari, M.; Mirniaharikandehei, S.; Khuzani, A.Z.; Danala, G.; Qiu, Y.; Zheng, B. Improving the performance of CNN to predict the likelihood of COVID-19 using chest X-ray images with preprocessing algorithms. Int. J. Med. Inform. 2020, 144, 104284. [Google Scholar] [CrossRef] [PubMed]
- Singh, D.; Singh, B. Investigating the impact of data normalization on classification performance. Appl. Soft Comput. 2020, 97, 105524. [Google Scholar] [CrossRef]
- Soares, E.; Angelov, P.; Biaso, S.; Froes, M.H.; Abe, D.K. SARS-CoV-2 CT-scan dataset: A large dataset of real patients CT scans for SARS-CoV-2 identification. MedRxiv 2020. [Google Scholar] [CrossRef]
- Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar]
- Sharma, M.; Verma, A.; Vig, L. Learning to clean: A GAN perspective. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 174–185. [Google Scholar]
- Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 694–711. [Google Scholar]
- Trapero, H.; Ilao, J.; Lacaza, R. An Integrated Theory for Chatbot Use in Air Travel: Questionnaire Development and Validation. In Proceedings of the 2020 IEEE REGION 10 CONFERENCE (TENCON), Osaka, Japan, 16–19 November 2020; pp. 652–657. [Google Scholar]
- Kuncheva, L.I.; Whitaker, C.J.; Shipp, C.A.; Duin, R.P. Limits on the majority vote accuracy in classifier fusion. Pattern Anal. Appl. 2003, 6, 22–31. [Google Scholar] [CrossRef]
- Choromanska, A.; Henaff, M.; Mathieu, M.; Arous, G.; LeCun, Y. The loss surfaces of multilayer networks in Artificial Intelligence and Statistics. PMLR 2015, 38, 192–204. [Google Scholar]
- Gonwirat, S.; Surinta, O. DeblurGAN-CNN: Effective Image Denoising and Recognition for Noisy Handwritten Characters. IEEE Access 2022, 10, 90133–90148. [Google Scholar] [CrossRef]
Figure 1.
Treatment strategy for tuberculosis (adapted from World Health Organization [
3]).
Figure 1.
Treatment strategy for tuberculosis (adapted from World Health Organization [
3]).
Figure 2.
The development of drug response classification system (DRCS) (DBOQS).
Figure 2.
The development of drug response classification system (DRCS) (DBOQS).
Figure 3.
Image augmentation example.
Figure 3.
Image augmentation example.
Figure 4.
Framework of the data normalizing model.
Figure 4.
Framework of the data normalizing model.
Figure 5.
A flow diagram to illustrate image pre-processing steps to generate the input of a CNN model.
Figure 5.
A flow diagram to illustrate image pre-processing steps to generate the input of a CNN model.
Figure 6.
Basic structure of MobileNetV2.
Figure 6.
Basic structure of MobileNetV2.
Figure 7.
Architecture of EfficientNetB1.
Figure 7.
Architecture of EfficientNetB1.
Figure 8.
Example of a deep DenseNet with three dense blocks.
Figure 8.
Example of a deep DenseNet with three dense blocks.
Figure 9.
The NASNetMobile architecture.
Figure 9.
The NASNetMobile architecture.
Figure 10.
Structure of the EDL using heterogeneous and homogenous manners.
Figure 10.
Structure of the EDL using heterogeneous and homogenous manners.
Figure 11.
Framework of how to use weight generation to find the optimal weight of ensemble DL.
Figure 11.
Framework of how to use weight generation to find the optimal weight of ensemble DL.
Figure 12.
The framework of the proposed model used to classify the drug response of the TB patients.
Figure 12.
The framework of the proposed model used to classify the drug response of the TB patients.
Figure 13.
The design framework of DBOQS for the case study.
Figure 13.
The design framework of DBOQS for the case study.
Figure 14.
Confusion matrices of the classification result shown in
Table 4.
Figure 14.
Confusion matrices of the classification result shown in
Table 4.
Figure 15.
Q and A results of DBOQS.
Figure 15.
Q and A results of DBOQS.
Table 1.
Number of data of each category.
Table 1.
Number of data of each category.
| DS-TB | DR-TB | MDR-TB | XDR-TB |
---|
Train set | 1299 | 375 | 1653 | 688 |
Test set | 308 | 93 | 434 | 169 |
Total | 1607 | 468 | 2087 | 857 |
Table 2.
Details of the compared method in the existing research.
Table 2.
Details of the compared method in the existing research.
Research | Classes | Features | Region in CXR | AUC | Accuracy F-Measure | Accuracy |
---|
Ureta and Shrestha [76] | DS vs. DR | CNN | Whole | 67.0 | - | - |
Tulo et al. [77] | DS vs. DR | Shape | Mediastinum + Lungs | | 93.6% | |
Jaeger et al. [24] | DS vs. MDR | Texture, Shape, and Edge | Lung | 66% | 61% | 62% |
Kovalev et al. [78] | DS vs. DR | Texture and Shape | Lung | - | - | 61.7 |
Tulo et al. [12] | DS vs. MDR | Shape | Mediastinum + Lungs | 87.3 | 82.4 | 82.5 |
Tulo et al. [12] | MDR vs. XDR | Shape | Mediastinum + Lungs | 86.6 | 81.0 | 81.0 |
Tulo et al. [12] | DS vs. XDR | Shape | Mediastinum + Lungs | 93.5 | 87.0 | 87.0 |
Karki et al. [11] | DS vs. DR | CNN | Lung excluded | 79.0 | - | 72.0 |
Proposed method | DR vs. DS | Ensemble CNN | Whole | - | - | - |
Proposed method | DS vs. MDR | Ensemble CNN | Whole | - | - | - |
Proposed method | DS vs. XDR | Ensemble CNN | Whole | - | - | - |
Proposed method | DR vs. MDR vs. XDR | Ensemble CNN | Whole | - | - | - |
Table 3.
Initial WP when D = 10 and i = 5.
Table 3.
Initial WP when D = 10 and i = 5.
NP\D | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|
1 | 0.64 | 0.82 | 0.61 | 0.93 | 0.40 | 0.99 | 0.70 | 0.68 | 0.36 | 0.08 |
2 | 0.58 | 0.28 | 0.99 | 0.79 | 0.26 | 0.60 | 0.09 | 0.20 | 0.09 | 0.85 |
3 | 0.48 | 0.90 | 0.04 | 0.95 | 0.85 | 0.29 | 0.09 | 0.08 | 0.93 | 0.49 |
4 | 0.34 | 0.69 | 0.91 | 0.12 | 0.33 | 0.97 | 0.20 | 0.73 | 0.14 | 0.61 |
5 | 0.88 | 0.25 | 0.28 | 0.32 | 0.51 | 0.75 | 0.06 | 0.73 | 0.10 | 0.93 |
Table 4.
Updated role of heuristics information.
Table 4.
Updated role of heuristics information.
Variables | Update Method |
---|
| Sum of work packages (WPs) that have selected IB b from iteration 1 through iteration t |
| The average objective value selected by IBs |
| When
|
| Update current global best WP |
| Update IB’s best WP |
| Pick a random positional value across all WPs |
Table 5.
Conclusions of 36 proposed models for drug response classification of TB patients.
Table 5.
Conclusions of 36 proposed models for drug response classification of TB patients.
Model | Preprocessing | CNN Architectures | Decision Fusion Strategy |
---|
Data Augmentation | Data Normalizing | EfficientNetB7 | DenseNet121 | NASNetMobile | MobileNetV2 | UMA | MV | AMIS-EDL |
---|
N-1 | | | ✓ | | | | | | |
N-2 | | | ✓ | | | | | ✓ | |
N-3 | | | ✓ | | | | | | ✓ |
N-4 | | | | ✓ | | | ✓ | | |
N-5 | | | | ✓ | | | | ✓ | |
N-6 | | | | ✓ | | | | | ✓ |
N-7 | | | | | ✓ | | ✓ | | |
N-8 | | | | | ✓ | | | ✓ | |
N-9 | | | | | ✓ | | | | ✓ |
N-10 | | | | | | ✓ | ✓ | | |
N-11 | | | | | | ✓ | | ✓ | |
N-12 | | | | | | ✓ | | | ✓ |
A-1 | ✓ | | ✓ | | | | ✓ | | |
A-2 | ✓ | | ✓ | | | | | ✓ | |
A-3 | ✓ | | ✓ | | | | | | ✓ |
A-4 | ✓ | | | ✓ | | | ✓ | | |
A-5 | ✓ | | | ✓ | | | | ✓ | |
A-6 | ✓ | | | ✓ | | | | | ✓ |
A-7 | ✓ | | | | ✓ | | ✓ | | |
A-8 | ✓ | | | | ✓ | | | ✓ | |
A-9 | ✓ | | | | ✓ | | | | ✓ |
A-10 | ✓ | | | | | ✓ | ✓ | | |
A-11 | ✓ | | | | | ✓ | | ✓ | |
A-12 | ✓ | | | | | ✓ | | | ✓ |
D-1 | | ✓ | ✓ | | | | ✓ | | |
D-2 | | ✓ | ✓ | | | | | ✓ | |
D-3 | | ✓ | ✓ | | | | | | ✓ |
D-4 | | ✓ | | ✓ | | | ✓ | | |
D-5 | | ✓ | | ✓ | | | | ✓ | |
D-6 | | ✓ | | ✓ | | | | | ✓ |
D-7 | | ✓ | | | ✓ | | ✓ | | |
D-8 | | ✓ | | | ✓ | | | ✓ | |
D-9 | | ✓ | | | ✓ | | | | ✓ |
D-10 | | ✓ | | | | ✓ | ✓ | | |
D-11 | | ✓ | | | | ✓ | | ✓ | |
D-12 | | ✓ | | | | ✓ | | | ✓ |
AD-1 | ✓ | ✓ | ✓ | | | | ✓ | | |
AD-2 | ✓ | ✓ | ✓ | | | | | ✓ | |
AD-3 | ✓ | ✓ | ✓ | | | | | | ✓ |
AD-4 | ✓ | ✓ | | ✓ | | | ✓ | | |
AD-5 | ✓ | ✓ | | ✓ | | | | ✓ | |
AD-6 | ✓ | ✓ | | ✓ | | | | | ✓ |
AD-7 | ✓ | ✓ | | | ✓ | | ✓ | | |
AD-8 | ✓ | ✓ | | | ✓ | | | ✓ | |
AD-9 | ✓ | ✓ | | | ✓ | | | | ✓ |
AD-10 | ✓ | ✓ | | | | ✓ | ✓ | | |
AD-11 | ✓ | ✓ | | | | ✓ | | ✓ | |
AD-12 | ✓ | ✓ | | | | ✓ | | | ✓ |
M-1 | | | ✓ | ✓ | ✓ | ✓ | ✓ | | |
M-2 | | | ✓ | ✓ | ✓ | ✓ | | ✓ | |
M-3 | | | ✓ | ✓ | ✓ | ✓ | | | ✓ |
M-4 | ✓ | | ✓ | ✓ | ✓ | ✓ | ✓ | | |
M-5 | ✓ | | ✓ | ✓ | ✓ | ✓ | | ✓ | |
M-6 | ✓ | | ✓ | ✓ | ✓ | ✓ | | | ✓ |
M-7 | | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | | |
M-8 | | ✓ | ✓ | ✓ | ✓ | ✓ | | ✓ | |
M-9 | | ✓ | ✓ | ✓ | ✓ | ✓ | | | ✓ |
M-10 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | | |
M-11 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | | ✓ | |
M-12 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | | | ✓ |
Table 6.
Multi-Class Classification Result.
Table 6.
Multi-Class Classification Result.
Methods | AUC | F-Measure | Accuracy | Methods | AUC | F-Measure | Accuracy |
---|
N-1 | 76.1 | 71.5 | 73.2 | D-7 | 80.7 | 75.5 | 77.0 |
N-2 | 75.9 | 72.1 | 74.7 | D-8 | 78.5 | 73.5 | 75.6 |
N-3 | 76.9 | 73.1 | 75.4 | D-9 | 83.4 | 78.2 | 80.4 |
N-4 | 76.4 | 70.5 | 71.4 | D-10 | 79.5 | 75.2 | 76.4 |
N-5 | 74.9 | 71.6 | 72.2 | D-11 | 79.5 | 76.5 | 78.4 |
N-6 | 77.4 | 73.8 | 75.6 | D-12 | 81.1 | 75.5 | 77.5 |
N-7 | 72.5 | 69.4 | 70.8 | AD-1 | 85.4 | 80.7 | 83.2 |
N-8 | 74.8 | 70.4 | 71.2 | AD-2 | 85.6 | 82.3 | 84.6 |
N-9 | 76.5 | 71.4 | 74.2 | AD-3 | 86.9 | 83.5 | 85.2 |
N-10 | 77.9 | 73.5 | 75.2 | AD-4 | 84.6 | 82.5 | 83.4 |
N-11 | 77.3 | 72.5 | 75.3 | AD-5 | 83.8 | 80.3 | 81.3 |
N-12 | 76.3 | 73.1 | 74.2 | AD-6 | 83.0 | 81.0 | 82.1 |
A-1 | 78.7 | 75.7 | 77.6 | AD-7 | 83.4 | 81.8 | 82.1 |
A-2 | 80.2 | 77.6 | 78.3 | AD-8 | 83.2 | 81.0 | 82.4 |
A-3 | 80.6 | 78.1 | 79.0 | AD-9 | 84.5 | 80.7 | 81.4 |
A-4 | 77.5 | 71.3 | 74.4 | AD-10 | 84.1 | 81.6 | 82.4 |
A-5 | 76.3 | 71.9 | 73.2 | AD-11 | 84.5 | 82.2 | 82.3 |
A-6 | 76.8 | 73.5 | 75.2 | AD-12 | 86.7 | 87.5 | 82.3 |
A-7 | 76.4 | 72.5 | 74.0 | M-1 | 84.5 | 82.3 | 83.5 |
A-8 | 77.8 | 73.5 | 74.5 | M-2 | 84.5 | 81.2 | 82.8 |
A-9 | 78.3 | 75.0 | 76.7 | M-3 | 84.8 | 83.2 | 84.6 |
A-10 | 78.4 | 74.9 | 75.3 | M-4 | 85.3 | 82.3 | 83.6 |
A-11 | 79.3 | 75.5 | 76.1 | M-5 | 86.5 | 83.5 | 84.5 |
A-12 | 78.9 | 75.0 | 76.6 | M-6 | 84.9 | 82.7 | 83.6 |
D-1 | 82.4 | 77.7 | 78.8 | M-7 | 86.5 | 84.1 | 85.4 |
D-2 | 80.2 | 77.9 | 79.0 | M-8 | 86.4 | 81.6 | 82.4 |
D-3 | 84.4 | 81.2 | 82.0 | M-9 | 87.1 | 85.2 | 86.6 |
D-4 | 80.1 | 77.8 | 79.7 | M-10 | 88.4 | 85.8 | 86.1 |
D-5 | 81.4 | 78.7 | 79.4 | M-11 | 88.5 | 85.5 | 86.2 |
D-6 | 82.5 | 79.4 | 80.6 | M-12 | 88.9 | 85.6 | 87.6 |
Table 7.
Percentage average accuracy value of using different strategies of the data presented in
Table 6.
Table 7.
Percentage average accuracy value of using different strategies of the data presented in
Table 6.
Preprocessing Techniques | CNN Architectures | Decision Fusion Strategies |
---|
No Preprocessing | Data Augmentation | Data Normalizing | Use All | EfficientNetB7 | DenseNet121 | NASNetMobile | MobileNetV2 | Use All | UMA | MV | AMIS-EDL |
---|
74.96 | 80.5 | 81.7 | 82.7 | 80.6 | 79.2 | 78.7 | 79.4 | 84.8 | 78.7 | 78.7 | 80.1 |
Table 8.
The accuracy of the proposed methods with the existing heuristics to classify DR-TB and DS-TB.
Table 8.
The accuracy of the proposed methods with the existing heuristics to classify DR-TB and DS-TB.
Methods | Classes | Features | Region in CXR | Accuracy |
---|
| | | | AUC | F-Measure | Accuracy |
---|
Ureta and Shrestha [76] | DS vs. DR | CNN | Whole | 67.0 | - | - |
Tulo et al. [77] | DS vs. DR | Shape | Mediastinum + Lungs | - | 93.6 | - |
Kovalev et al. [78] | DS vs. DR | Texture and Shape | Lung | - | - | 61.7 |
Karki et al. [11] | DS vs. DR | CNN | Lung excluded | 79.0 | - | 72.0 |
Proposed method | DS vs. DR | Ensemble CNN | Whole | 97.9 | 93.8 | 94.8 |
Table 9.
The accuracy of the proposed methods with the existing heuristics to classify DS-TB and MDR.
Table 9.
The accuracy of the proposed methods with the existing heuristics to classify DS-TB and MDR.
Methods | Classes | Features | Region in CXR | | Accuracy | |
---|
| | | | AUC | F-Measure | Accuracy |
---|
Jaeger et al. [24] | DS vs. MDR | Texture, Shape, and Edge | Lung | 66 | 61 | 62 |
Tulo et al. [12] | DS vs. MDR | Shape | Mediastinum + Lungs | 87.3 | 82.4 | 82.5 |
Proposed method | DS vs. MDR | Ensemble CNN | Whole | 89.5 | 88.0 | 88.5 |
Table 10.
The accuracy of the proposed methods with the existing heuristics to classify DS-TB and XDR.
Table 10.
The accuracy of the proposed methods with the existing heuristics to classify DS-TB and XDR.
| Classes | Features | Region in CXR | | Accuracy | |
---|
| | | | AUC | F-Measure | Accuracy |
---|
Tulo et al. [12] | DS vs. XDR | Shape | Mediastinum + Lungs | 93.5 | 87.0 | 87.0 |
Proposed method | DS vs. XDR | Ensemble CNN | Whole | 95.1 | 88.7 | 89.1 |
Table 11.
The accuracy of the proposed methods with the existing heuristics to classify MDR and XDR.
Table 11.
The accuracy of the proposed methods with the existing heuristics to classify MDR and XDR.
| Classes | Features | Region in CXR | | Accuracy | |
---|
| | | | AUC | F-Measure | Accuracy |
---|
Tulo et al. [12] | MDR vs. XDR | Shape | Mediastinum + Lungs | 86.6 | 81.0 | 81.0 |
Proposed method | MDR vs. XDR | Ensemble CNN | Whole | 88.7 | 82.5 | 84.9 |
Table 12.
The DBOQS accuracy results.
Table 12.
The DBOQS accuracy results.
| Number of Input Images | Number of Correct Classification | %Correct Classification | Number of Wrong Classification | %Wrong Classification |
---|
DS | 104 | 94 | 90.38 | 10 | 9.62 |
DR | 121 | 109 | 90.08 | 12 | 9.92 |
MDR | 115 | 105 | 91.30 | 10 | 8.70 |
XDR | 103 | 94 | 91.26 | 9 | 8.74 |
Total | 443 | 402 | 90.74 | 41 | 9.26 |
Table 13.
Questionnaire result from 30 doctors who were trying to use DRCS.
Table 13.
Questionnaire result from 30 doctors who were trying to use DRCS.
Questionnaire on Images | Level of Agreement (Strong Agreement Level Is 10) |
---|
DBOQS has fast responses. | 9.51 |
DBOQS can classify the drug response correctly. | 9.58 |
DBOQS may simplify the drug response diagnosis process. | 9.97 |
DBOQS can reduce your workload. | 9.86 |
You will use the DBOQS to aid in your drug response diagnosis. | 9.74 |
What score will you assign the DBOQS with regard to your trust level? | 9.51 |
What score will you assign the DBOQS with regard to your preferences? | 9.97 |
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).