Next Article in Journal
Waymarking in Social Robots: Environment Signaling Using Human–Robot Interaction
Next Article in Special Issue
Novel Privacy Preserving Non-Invasive Sensing-Based Diagnoses of Pneumonia Disease Leveraging Deep Network Model
Previous Article in Journal
A Non-Invasive Millimetre-Wave Radar Sensor for Automated Behavioural Tracking in Precision Farming—Application to Sheep Husbandry
Previous Article in Special Issue
Memcached: An Experimental Study of DDoS Attacks for the Wellbeing of IoT Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Transfer Learning for Classification of Skin Cancer

1
Department of Information Technology and Engineering, Vellore Institute of Technology, Vellore 632014, Tamil Nadu, India
2
Department of Computer Science and Engineering, Vellore Institute of Technology, Vellore 632014, Tamil Nadu, India
3
Industrial Engineering Department, College of Engineering, King Saud University, Riyadh 11421, Saudi Arabia
4
Industrial Engineering Department, College of Engineering, University of Houston, Houston, TX 77204-4008, USA
*
Author to whom correspondence should be addressed.
Submission received: 1 November 2021 / Revised: 30 November 2021 / Accepted: 30 November 2021 / Published: 6 December 2021

Abstract

:
One of the major health concerns for human society is skin cancer. When the pigments producing skin color turn carcinogenic, this disease gets contracted. A skin cancer diagnosis is a challenging process for dermatologists as many skin cancer pigments may appear similar in appearance. Hence, early detection of lesions (which form the base of skin cancer) is definitely critical and useful to completely cure the patients suffering from skin cancer. Significant progress has been made in developing automated tools for the diagnosis of skin cancer to assist dermatologists. The worldwide acceptance of artificial intelligence-supported tools has permitted usage of the enormous collection of images of lesions and benevolent sores approved by histopathology. This paper performs a comparative analysis of six different transfer learning nets for multi-class skin cancer classification by taking the HAM10000 dataset. We used replication of images of classes with low frequencies to counter the imbalance in the dataset. The transfer learning nets that were used in the analysis were VGG19, InceptionV3, InceptionResNetV2, ResNet50, Xception, and MobileNet. Results demonstrate that replication is suitable for this task, achieving high classification accuracies and F-measures with lower false negatives. It is inferred that Xception Net outperforms the rest of the transfer learning nets used for the study, with an accuracy of 90.48. It also has the highest recall, precision, and F-Measure values.

1. Introduction

Skin cancer is one of the most common cancers worldwide. It greatly affects the quality of life. The most common cause is the over exposure of skin to ultraviolet radiations coming from the sun [1]. The rate of being affected when exposed to UV radiations is higher in fair skinned, more sun-sensitive people than in dark skinned, less sun-sensitive people [2].
Invasive melanoma represents about 1% of all skin cancer cases, but it contributes to the majority of deaths in skin cancer. Incidence of melanoma skin cancer has risen rapidly over the past 30 years. It is estimated that in 2021, 100,350 new cases of melanoma will be diagnosed in the US and around 6850 people will eventually die from it [3].
The best way to control skin cancer is its early detection and prevention [4]. Awareness of new or changing skin spots or growths, particularly those that look unusual, should be evaluated. Any new lesions, or progressive change in a lesion’s appearance (size, shape, or color), should be evaluated by a clinician. With the advent of deep learning concepts [5], we can classify skin cancer detection in seven diagnostic categories, namely melanocytic nevi, melanoma, benign keratosis-like lesions, basal cell carcinoma, actinic keratosis, vascular lesions, and dermatofibroma. Generally, a dermatologist specializing in skin cancer detection follows a fixed sequence, i.e., starting with a visual examination of the suspected lesion with naked eyes, followed by a dermoscopy and finally a biopsy [6].
In today’s era, with the usage of artificial intelligence and deep learning [7] in medical diagnostics [8], the efficiency of predicting a result increases exponentially as compared to the dependency on a visual diagnostic [9,10,11], Machine learning also has applications in many other fields, alongside the medical field [12,13,14,15,16]. The convolutional neural network (CNN) is an important artificial intelligence algorithm in feature selection and object classification [17,18,19]. Deep convolutional neural networks (DCNN) help in classifying skin lesions into seven different categories, with the help of their dermoscopic images, covering all the lesions found in skin cancer identification. Although DNNs require a large amount of data for training [20,21], they have an appealing impact on medical image classification [22,23]. DNNs train a network of large-scale datasets using high performance GPUs, thus providing a better outcome [17,24]. Deep learning algorithms backed by these high performing GPUs in computing large datasets have shown better performance than humans in skin cancer identification [25].

Literature Background

Deep learning gained popularity during the last decade [26,27,28]. Convolutional neural networks have been widely used in the classification of diseases [29,30]. It is challenging to train a CNN architecture if the datasets have a limited number of training samples. In [18], a partial transferable CNN was proposed in order to cope with a new dataset with a different spatial resolution, a different number of bands, and variation in the signal-to-noise ratio. The experimental results using different state-of-the-art models show that partial CNN transfer with even-numbered layers provides better mapping accuracy for the target dataset with a limited number of training samples. In [19], a novel method using transfer learning to deal with multi-resolution images from various sensors via CNN is proposed. CNN trained for a typical image data set, and the trained weights were transferred to other data sets of different resolutions. Initially, skin cancer diseases were divided only into two categories, benign or malignant. Canziani et al. [31] made use of machine learning algorithms, such as K-Means and SVM, and achieved an accuracy of 90%. Codella [32] makes use of the ISIC 2017 dataset, which consists of three categories of skin cancer, with conventional machine learning methods, in order to predict melanoma precisely but suffered from inaccurate results due to dataset bias and incomplete dermoscopic feature annotations. Another case of skin lesion classification [33] on the same dataset, in which a proposed lesion indexing network (LIN) was introduced, managed to attain the 91.2% area under the curve. However, it was performed on ISIC 2017, and no work has been recreated on ISIC 2018. There are also some datasets that divide the skin lesion into 12 different categories. Han [34] used the Asan dataset, med-node dataset, and atlas site images, which, together, consisted of 19,398 images divided into 12 categories. He used Resnet architecture for classification and achieved an accuracy of 83%. His paper was moreover inclined towards proving that the proposed dataset was better than those taken in comparison. Chaturvedi et al. [35] made use of the HAM10000 dataset, seven different types of skin lesion, using MobileNet in skin lesion detection, and achieved an accuracy of 83%. Milton [36] presented with transfer learning algorithms that were trained on the HAM10000 dataset and used fine-tuning and freezing of two epochs. Here, PNASNet-5-Large was used, which gave an accuracy of 76%. HAM10000 being an unbalanced dataset with a large difference in total images for each class makes it harder to generalize the features of the lesions. Nugroho [25] made his own custom CNN model, which produced 78% accuracy on the HAM10000 dataset. Kadampur [5] introduced an online method without coding for the classification of HAM10000 diseases and training on the cloud. Although the advantage of the above-mentioned research works is that they provide a straightforward algorithm approach and acceptable accuracy, most of them did not consider all types of legions and used relatively old datasets.
It was found that most papers had done classification of lesions [37] in the three standard categories, i.e., basal cell carcinoma, squamous cell carcinoma, and melanoma. The dataset used for classification was not so recent and not sufficient enough to identify all types of lesions [38]. By keeping all this in mind, three objectives were framed
  • To classify the images from HAM10000 dataset into seven different types of skin cancer.
  • To use transfer learning nets for feature selection and classification so as to identify all types of lesions found in skin cancer.
  • To properly balance the dataset using replication on only training data and perform a detailed analysis using different transfer learning models.
In this paper, HAM10000 dataset was used to train the model for skin cancer classification. All the six transfer learning nets were compared, and their training and validation loss, training and validation accuracy, along with their individual confusion matrices, were plotted. A comparative analysis of accuracy was then performed for all these learning nets and concluded with the model, which gave the highest accuracy in identifying all the lesions.

2. Materials and Methods

2.1. Dataset Description for Skin Lesion

To carry out the research work, we used HAM10000 dataset (human against machine) [39], which has 10015 dermatoscopic images and seven different classes, such as actinic keratosis (akiec) (327), basal cell carcinoma (bcc) (541), benign keratosis (bkl) (1099), dermatofibroma (df) (155), melanocytic nevi (nv) (6705), melanoma (mel) (1113), and vascular skin lesions (vasc) (142). Seven types of lesions [29] are shown in Figure 1, along with their occurrences in Figure 2, where x-axis represents the type of lesion and y-axis represents the corresponding count. The same dataset was divided into training, testing, and validation sets, so that there was no discrepancy in the results.

2.2. Transfer Learning Nets

In this section, the focus is on transfer learning and the models used in the research are discussed briefly [40]. Transfer learning is a machine learning method in which a model developed from one task is reused in another. It is generally used when we do not have enough training data. However, the data issue can be solved with data augmentation. The main reason why we need transfer learning is because Melanoma and benign lesions have high similarity, so it takes a long time to identify and classify them. Furthermore, transfer learning is more efficient in classifying between similar lesions, making it a first choice. Transfer learning nets are trained on large datasets and their model weights are frozen, and the last few layers are changed for a different dataset. In this paper, the models we used for comparison were VGG19, InceptionV3, InceptionResNetV2, Resnet50, Xception. and MobileNet. However, here, we not only used the frozen weights, but we also retrained them on our dataset so that the network layers had better precision in distinguishing between seven different types of lesions. We trained the models on the skin lesion dataset using these six transfer learning nets and analyzed their predictions. In addition to this, we plotted their training and validation loss, training and validation accuracy, along with their individual confusion matrices. A comparative analysis of accuracy of all these learning nets was then performed, and we determined the net that gave the highest accuracy in identifying all the lesions.

2.2.1. VGG19

This network is characterized by its simplicity. It has five blocks each of 3 × 3 convolutional layers stacked on top of each other. Volume size is reduced by max pooling of 2 × 2 kernels and a stride of 2. It is followed by two fully-connected layers, each of 4096 nodes, with ReLU activation function. The final layer has 1000 nodes with softmax as its activation function [41]. VGG19 has about 143 million parameters in total. Some applications of VGG net are mentioned by Canziani [31] in his paper.

2.2.2. InceptionV3

InceptionV3 [42] is the refined version of the GoogLeNet architecture [43]. The basic idea of this net is to make this process simpler and more efficient. The Inception module acts as a multi-level feature extractor. It computes 1 × 1, 3 × 3, and 5 × 5 convolutions within the same module of the network. The outputs of these filters are then stacked on each other and fed into the next layer in the network.

2.2.3. InceptionResnetv2

In this net [44], the residual version of Inception nets is used rather than simple inception modules. Each Inception block is followed by a filter-expansion layer (1 × 1 convolution without activation), which is used for scaling up the dimensionality of the filter bank before the addition to match the depth of the input. Inception-ResNet-v2 matches the computational cost of the Inception-v4 network. The difference between residual and non-residual Inception variants is that in the case of Inception-ResNetv2; batch normalization is used only on top of the traditional layers, but not on top of the summations.

2.2.4. ResNet50

These are the deeper convolutional neural nets, which make use of skip connections [45]. These residual blocks greatly resolve gradient degradation and also reduce total parameters. Residual Networks (ResNet [46]) architecture follows two simple design rules. Firstly, for the same output map size, layers have the same number of filters, and secondly, when the feature map size is halved, the filters count is doubled. Batch normalization is performed after each convolution layer and before ReLU activation function. If the input and output have the same size, the shortcut is used. When there is an increase in dimensions, the projection shortcut is used.

2.2.5. Xception

The Xception [47] architecture is an extension of the Inception architecture. It replaces the standard Inception modules with depth wise separable convolutions. It does not perform partitioning on input data and maps the spatial correlations for each output channel separately. The Xception net then performs 1 × 1 depth wise convolution, which captures cross-channel correlation. It slightly outperforms Inception V3 in terms of smaller data and vastly on bigger data.

2.2.6. MobileNet

This net makes use of depth wise separable connections, similar to the Xception net. For MobileNets [48], the depth wise convolution applies a single filter to each input channel. The pointwise convolution then applies a 1 × 1 convolution to combine the outputs of the depth wise convolution. A standard convolution layer does both: filters and combines inputs into a new set of outputs in one step. The depth wise separable convolution splits this into two layers: a separate layer for filtering and a separate layer for combining. This factorization has the effect of drastically reducing computation and model size. MobileNet is particularly useful for mobile and embedded vision applications. It has less parameters compared to others and also less complexity. This architecture is a concise form of the Xception and Inception nets.

2.3. Proposed Methodology

In this section, we explain the process of classification of skin lesions. The main issue with the dataset is that it is highly imbalanced and contains a lot of duplicated images. Therefore, we made use of data augmentation to resolve this issue. Figure 3 provides a diagrammatic representation of the proposed method.

2.3.1. Data Augmentation

When inspecting the dataset, it was observed that a lot of images were just a replication of each other, which is not beneficial for our models. We identified the unique images in the dataset, which amounted to be around 5514. We split these images into training and testing data of 80% and 20%, respectively. Training data was further divided into 90% training and 10% validation. Training data had approximately 4000 images, which was very few, and also, the classes were unbalanced because a few classes had many more images compared to others. After removing duplicates, Melanocytic Nevi had 3179 images and Dermatofibroma had only 28 images. We tackled this problem by replicating the class with low data by multiplying it by a factor that would produce data close to the class with the highest data.
Table 1 explains the frequency of images of each label before and after augmentation. All the images were multiplied by a factor k so that they could lie closest to Melanocytic Nevi. The above technique was used to avoid the problem of class imbalance. Here, we expanded the training dataset artificially. We altered the training data with small transformations to reproduce variations. A few techniques, such as rotation, zooming, and shifting vertically and horizontally, were implemented. All the images were part of the HAM10000 dataset, and their dimensions were resized from 450 × 600 to 128 × 128 dimensions for convenience in processing.
After training the model with the initial training dataset, it was observed that even though a good accuracy was obtained, an observation of the classification matrix reflected the real picture. One class (Melanocytic Nevi) was classified a majority of the time, since it had the highest frequency. This indicated that the model was biased and was not able to predict or classify other low frequency classes. To overcome this situation, it was required to equalize the distribution of classes and let the model understand each class. The dataset was augmented by increasing the frequency of each class so that all the classes had the same number of images. As a result, a better performing wholesome model could be realized.

2.3.2. Preprocessing

After the image acquisition task, we performed image preprocessing. We had three channels of data corresponding to the colors Red, Green, and Blue (RGB). Pixel levels are usually [0–255]. Image preprocessing involves normalization of the images. In normalization, mean and standard deviation of all images in the dataset is calculated. The mean of all the images was subtracted from initial images and then the obtained result was divided by standard deviation. On the other hand, the seven diseases were one, hot encoded, i.e., a binary column for each category was created.
Image width, image height = 128, 128
3, channels = pixel levels in the range [0–255]
Normalization = x μ σ
where x is the original feature vector, μ is the mean, and σ is the standard deviation.

2.3.3. Feature Extraction

Feature extraction is the most crucial step in classification. Feature extraction was carried out by pre-trained transfer learning models. This involves looking up important features in an image and then deriving information from them. Several CNNs are stacked up back-to-back in order to make a model.
Here, we used pre-trained models, such as VGG19, InceptionV3, Resnet50, Xception, InceptionResNetV2, and MobileNet. All of the above pre-trained nets used the weights of the Imagenet. The bottom layers were Max Pooling, which calculates the maximum value of each patch of feature map, Flatten, which converts the 3d array into a 1d array, Dense layer with 128 neutrons and finally a Dense layer with seven neurons, corresponding to seven different diseases with sigmoidal activation function.

2.3.4. Classification and Evaluation

The final layer outputs an array of seven values, which indicates the probability of each category of disease. The class number was in correspondence to seven different skin cancers. The class numbers assigned for different lesions were actinic keratosis (0), basal cell carcinoma (1), benign keratosis like lesions (2), dermatofibroma (3), melanocytic nevi (4), melanoma (5), and vascular skin lesions (6), and in the evaluation phase, we used a validation dataset for validating the different nets for the skin lesion dataset.

3. Results

In this section, the experimental results and analysis of our models used on the HAM10000 dataset are presented. The results of six different types of transfer learning models, VGG19, InceptionV3, Resnet50, Xception, InceptionResNetV2, and MobileNet, on the dataset with and without repetition of images were compared. All the models were trained for 10 epochs with a batch size of 32. In every epoch training accuracy, training error, and validation accuracy, validation error was calculated. We adopted an Adaptive Momentum (Adam) optimizer with a learning rate (LR) of 0.001 and a loss function as a Categorical Cross Entropy. In order to make the optimizer converge faster and closer to the global minimum, an annealing method was used with a LR. To keep the advantage of the faster computation time with a high LR, we decreased the LR dynamically every four epochs, depending on the validation accuracy. With the ‘ReduceLROnPlateau’ function from ‘Keras.callbacks’, we chose to reduce the LR by half if the validation loss did not improve after four epochs.
Table 2 and Table 3 show the results of the transfer learning nets’ accuracy with and without repetition of images.
Table 3 clearly shows the difference produced when images were equalized in their frequencies. While some nets, such as VGG19, InceptionV3, and ResNet50, showed a decline in accuracy, there was an increase in accuracy for InceptionResNetV2, MobileNet, and Xception. The best performing net in the balanced dataset was Xception, while in unbalanced dataset it was MobileNet.
Here, we compared the performance of all seven nets by using a confusion matrix. A confusion matrix was constructed for every network. Performance of different nets was tested by passing 1002 randomly selected images (testing data). Accuracy was considered as a measure for calculating the performance with the skin lesion dataset. Figure 4a–f shows the confusion matrix results that we achieved for different nets.
From Figure 4, we can infer true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values for each class. These values can help in calculating Precision, Recall, F1 Score, and accuracy.
Incorrectly classified graphs in Figure 5 are derived from the confusion matrices. Figure 5 shows the incorrectly classified results for different nets. They can be calculated as the percentage of incorrect classification out of the total number of images in each class. We can follow the same process for correctly classified graphs. Figure 5a–f shows the fraction that has been classified incorrectly by the different transfer learning nets and our proposed model. From all the subfigures, it is evidently clear that ‘Akiec’ (label ‘0’) and ‘Mel’ (label ‘5’) are the most difficult lesions to be identified in the system. Additionally, ‘Nv’ (label ‘4’), having the least incorrect classification, was correctly classified by the system. In the VGG19 model, ‘Bkl’ (label ‘2’) performed worst out of all the remaining classes. MobileNet presented the highest classification accuracy for label ‘2’, with respect to models. ‘Df’ (label ‘3’) was easily identifiable by VGG19 and InceptionV3 compared to the rest, whereas ‘Nv’ (label ‘4’) showed better results in Inception-ResNet-v2, MobileNet, and XceptionNet. From the analysis, Figure 5e has most of the lowest incorrectly classified values and the highest accuracy.
Figure 6a–f shows the training: validation accuracy and loss for all the six models. Figure 6a shows that VGG19 has a validation loss of 0.88 and an accuracy of 67.54% on skin lesion classification. Figure 6b shows that InceptionV3 has a validation loss of 0.66 and an accuracy of 86.40%, while Figure 6c shows that InceptionResNetV2 has a validation loss of 0.68 and an accuracy of 88.40%. Figure 6d shows that Xception has a validation loss of 0.58 and an accuracy of 89.66%, and Figure 6e shows that Resnet50 has a validation loss of 0.74 and an accuracy of 82.32%. Finally, Figure 6f shows that MobileNet has a validation loss of 0.65 and an accuracy of 87.21%. From the Figure 6a–f, we can also infer that Xception Net’s loss lies in the range of 0 to 2, thereby incurring minimum fluctuations.
P r e c i s i o n = T P T P + F P ,   R e c a l l = T P T P + F N
F 1 S c o r e = 2 * P r e c i s i o n * R e c a l l P r e c i s i o n + R e c a l l
A c c u r a c y = T P + T N T P + T N + F P + F N
where TP, TN, FP, and FN denote the true positives, true negatives, false positives, and false negatives, respectively. By making use of the above formulae, we can calculate precision, recall, and f-measure for each class and also get the information of which network performs best. In addition, the average of the recall, precision, and F-measure among classes will be computed in order to obtain a precise measure that it is not corrupted by the class imbalance. Note that the average recall is the equivalent to the balanced accuracy for multi-class problems. In the table below, we present the average values of precision, recall, F-measure, and accuracies.
From Table 4, it is clear that the Xception Net has the highest values in terms of accuracy, recall, precision, and F-Measure. Let’s take a look at the individual values of each class in the Xception Net to get a better understanding.
From the above Table 5, Label 4 (Nv) has the highest precision of 94% and Label 5 (Mel) has the lowest, with 50%. Label 6 (vasc), although having a lower number of images in the validation set, has proved to show a good precision, along with Label 0 (Akiec) and Label 1 (Bcc), respectively. Table 6 gives the results of accuracies and losses on our test set. We infered that the Xception net had the lowest loss of 0.5168 and the highest achieved accuracy of 90.48% in our work.

Computational Cost

The computational costs of the simulations is provided here (Table 7 and Table 8) in the form of hardware used and the computational cost involved for different models.

4. Discussion and Conclusions

Due to the COVID-19 situation, everyone has suffered a lot but also gained a lot [49]. On one side, a large number of populations have contracted coronavirus, and many have died, but it is nowhere near to the upcoming UV radiations, which would have penetrated the ozone layer. Because of this pandemic situation and people staying in their homes, this has caused the ozone layer hole, which was getting bigger day by day, to close up. Skin cancers can be now diagnosed using these tools and can be treated earlier, and we can save more lives.
Through this research work, it is demonstrated that it is possible to achieve a competitive classification performance by using different types of data augmentation and transfer learning methods. Using the data augmentation method, we could get nearly 32k images, and we then performed feature extraction to get the required results. It is inferred that Xception Net outperforms the rest of the transfer learning nets used. It was observed that Label 0 (Akiec) and Label 5 (Mel) were most incorrectly classified because of their extreme resemblance to simple skin patches that are not harmful. Xception Net provides us with an accuracy of 90.48. It has the highest recall, precision, and F-Measure values, which are 89.57, 88.76, and 89.02 respectively. InceptionResNetV2 and MobileNet follow Xception Net closely in the prediction of results. Melanocytic Nevi is the most accurately classified skin lesion. There is a necessity of further and extensive research in this field as skin cancer-caused deaths are taking a toll. Transfer learning algorithms differ from those used in this paper, and proper fine-tuning may result in better accuracy.

Author Contributions

Conceptualization, S.J., U.S., B.T. and E.A.N.; methodology, S.J., U.S., B.T. and M.K.A.; software, S.J., U.S., B.T. and E.A.N.; formal analysis, S.J., U.S. and B.T.; resources, B.T., E.A.N., M.K.A. and A.K.K.; writing—original draft preparation, S.J., U.S., B.T. and E.A.N.; writing—review and editing, M.K.A. and A.K.K.; supervision, B.T., E.A.N. and A.K.K.; funding acquisition, E.A.N. All authors have read and agreed to the published version of the manuscript.

Funding

King Saud University for funding this work through Researchers Supporting Project number (RSP-2021/164), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available in the manuscript.

Acknowledgments

The authors extend their appreciation to King Saud University for funding this work through Researchers Supporting Project number (RSP-2021/164), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kricker, A.; Armstrong, B.K.; English, D.R. Sun exposure and non-melanocytic skin cancer. Cancer Causes Control 1994, 5, 367–392. [Google Scholar] [CrossRef] [PubMed]
  2. Armstrong, B.K.; Kricker, A. The epidemiology of UV induced skin cancer. J. Photochem. Photobiol. B 2001, 63, 8–18. [Google Scholar] [CrossRef]
  3. American Cancer Society. Cancer Facts and Figures. 2020. Available online: https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/annual-cancer-facts-and-figures/2020/cancer-facts-and-figures-2020.pdf (accessed on 5 July 2021).
  4. Silberstein, L.; Anastasi, J. Hematology: Basic Principles and Practice; Elsevier: Amsterdam, The Netherlands, 2017; p. 2408. [Google Scholar]
  5. Kadampur, M.A.; Al Riyaee, S. Skin cancer detection: Applying a deep learning based model driven architecture in the cloud for classifying dermal cell images. Inform. Med. Unlocked 2020, 18, 100282. [Google Scholar] [CrossRef]
  6. Haenssle, H.A.; Fink, C.; Schneiderbauer, R.; Toberer, F.; Buhl, T.; Blum, A.; Kalloo, A.; Hassen, A.B.H.; Thomas, L.; Enk, A.; et al. Man against machine: Diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. 2018, 29, 1836–1842. [Google Scholar] [CrossRef] [PubMed]
  7. Bhattacharyya, S.; Snasel, V.; Hassanien, A.E.; Saha, S.; Tripathy, B.K. Deep Learning: Research and Applications; De Gruyter: Berlin, Germany; Boston, MA, USA, 2020. [Google Scholar] [CrossRef]
  8. Dhanamjayulu, C.; Nizhal, N.U.; Maddikunta, P.K.R.; Gadekallu, T.R.; Iwendi, C.; Wei, C.; Xin, Q. Identification of malnutrition and prediction of BMI from facial images using real-time image processing and machine learning. IET Image Process. 2021. [Google Scholar] [CrossRef]
  9. Adate, A.; Arya, D.; Shaha, A.; Tripathy, B.K. Deep Learning: Research and Applications. In 4 Impact of Deep Neural Learning on Artificial Intelligence Research; Siddhartha, B., Vaclav, S., Aboul Ella, H., Satadal, S., Tripathy, B.K., Eds.; De Gruyter: Berlin, Germany, 2020; pp. 69–84. [Google Scholar] [CrossRef]
  10. Adate, A.; Tripathy, B.K. Deep Learning Techniques for Image Processing Machine Learning for Big Data Analysis; Bhattacharyya, S., Bhaumik, H., Mukherjee, A., De, S., Eds.; De Gruyter: Berlin, Germany, 2018; pp. 69–90. [Google Scholar] [CrossRef]
  11. Ayoub, A.; Mahboob, K.; Javed, A.R.; Rizwan, M.; Gadekallu, T.R.; Alkahtani, M.; Abidi, M.H. Classification and Categorization of COVID-19 Outbreak in Pakistan. Comput. Mater. Contin. 2021, 69, 1253–1269. [Google Scholar] [CrossRef]
  12. Abidi, M.H.; Alkhalefah, H.; Mohammed, M.K.; Umer, U.; Qudeiri, J.E.A. Optimal Scheduling of Flexible Manufacturing System Using Improved Lion-Based Hybrid Machine Learning Approach. IEEE Access 2020, 8, 96088–96114. [Google Scholar] [CrossRef]
  13. Abidi, M.H.; Umer, U.; Mohammed, M.K.; Aboudaif, M.K.; Alkhalefah, H. Automated Maintenance Data Classification Using Recurrent Neural Network: Enhancement by Spotted Hyena-Based Whale Optimization. Mathematics 2020, 8, 2008. [Google Scholar] [CrossRef]
  14. Ch, R.; Gadekallu, T.R.; Abidi, M.H.; Al-Ahmari, A. Computational System to Classify Cyber Crime Offenses using Machine Learning. Sustainability 2020, 12, 4087. [Google Scholar] [CrossRef]
  15. Abidi, M.H.; Alkhalefah, H.; Umer, U. Fuzzy harmony search based optimal control strategy for wireless cyber physical system with industry 4.0. J. Intell. Manuf. 2021. [Google Scholar] [CrossRef]
  16. Kumar, P.A.; Shankar, G.S.; Maddikunta, P.K.R.; Gadekallu, T.R.; Al-Ahmari, A.; Abidi, M.H. Location Based Business Recommendation Using Spatial Demand. Sustainability 2020, 12, 4124. [Google Scholar]
  17. Marks, R. The epidemiology of non-melanoma skin cancer: Who, why and what can we do about it. J. Dermatol. 1995, 22, 853–857. [Google Scholar] [CrossRef]
  18. Farooq, A.; Jia, X.; Hu, J.; Zhou, J. Transferable Convolutional Neural Network for Weed Mapping With Multisensor Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 1–16. [Google Scholar] [CrossRef]
  19. Farooq, A.; Jia, X.; Hu, J.; Zhou, J. Knowledge Transfer via Convolution Neural Networks for Multi-Resolution Lawn Weed Classification. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  20. Bose, A.; Tripathy, B.K. Deep Learning for Audio Signal Classification. In Deep Learning: Research and Applications; Bhattacharyya, S., Snasel, V., Ella Hassanien, A., Saha, S., Tripathy, B.K., Eds.; De Gruyter: Berlin, Germany, 2020; pp. 105–136. [Google Scholar] [CrossRef]
  21. Ghayvat, H.; Pandya, S.N.; Bhattacharya, P.; Zuhair, M.; Rashid, M.; Hakak, S.; Dev, K. CP-BDHCA: Blockchain-based Confidentiality-Privacy preserving Big Data scheme for healthcare clouds and applications. IEEE J. Biomed. Health Inform. 2021, 1. [Google Scholar] [CrossRef]
  22. Shah, A.; Ahirrao, S.; Pandya, S.; Kotecha, K.; Rathod, S. Smart Cardiac Framework for an Early Detection of Cardiac Arrest Condition and Risk. Front. Public Health 2021, 9, 762303. [Google Scholar] [CrossRef] [PubMed]
  23. Ghayvat, H.; Awais, M.; Gope, P.; Pandya, S.; Majumdar, S. ReCognizing SUspect and PredictiNg ThE SpRead of Contagion Based on Mobile Phone LoCation DaTa (COUNTERACT): A system of identifying COVID-19 infectious and hazardous sites, detecting disease outbreaks based on the internet of things, edge computing, and artificial intelligence. Sustain. Cities Soc. 2021, 69, 102798. [Google Scholar] [CrossRef]
  24. Kaul, D.; Raju, H.; Tripathy, B.K. Deep Learning in Healthcare. In Deep Learning in Data Analytics: Recent Techniques, Practices and Applications; Acharjya, D.P., Mitra, A., Zaman, N., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 97–115. [Google Scholar] [CrossRef]
  25. Nugroho, A.A.; Slamet, I.; Sugiyanto. Skins cancer identification system of HAMl0000 skin cancer dataset using convolutional neural network. AIP Conf. Proc. 2019, 2202, 020039. [Google Scholar] [CrossRef]
  26. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep Learning for Computer Vision: A Brief Review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef] [PubMed]
  27. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  28. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A Review on Deep Learning Techniques Applied to Semantic Segmentation. arXiv 2017, arXiv:1704.06857. Available online: https://ui.adsabs.harvard.edu/abs/2017arXiv170406857G (accessed on 1 April 2017).
  29. Maheshwari, K.; Shaha, A.; Arya, D.; Rajasekaran, R.; Tripathy, B.K. Deep Learning: Research and Applications. In Convolutional Neural Networks: A Bottom-Up Approach; Siddhartha, B., Vaclav, S., Aboul Ella, H., Satadal, S., Tripathy, B.K., Eds.; De Gruyter: Berlin, Germany, 2020; pp. 21–50. [Google Scholar] [CrossRef]
  30. Mateen, M.; Wen, J.; Nasrullah; Song, S.; Huang, Z. Fundus Image Classification Using VGG-19 Architecture with PCA and SVD. Symmetry 2019, 11, 1. [Google Scholar] [CrossRef] [Green Version]
  31. Canziani, A.; Paszke, A.; Culurciello, E. An Analysis of Deep Neural Network Models for Practical Applications. arXiv 2016, arXiv:1605.07678. [Google Scholar]
  32. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 International symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
  33. Li, Y.; Shen, L. Skin Lesion Analysis towards Melanoma Detection Using Deep Learning Network. Sensors 2018, 18, 556. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Han, S.S.; Kim, M.S.; Lim, W.; Park, G.H.; Park, I.; Chang, S.E. Classification of the Clinical Images for Benign and Malignant Cutaneous Tumors Using a Deep Learning Algorithm. J. Investig. Dermatol. 2018, 138, 1529–1538. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Chaturvedi, S.S.; Gupta, K.; Prasad, P.S. Skin Lesion Analyser: An Efficient Seven-Way Multi-class Skin Cancer Classification Using MobileNet. Adv. Mach. Learn. Technol. Appl. 2020, 1141, 165–176. [Google Scholar] [CrossRef]
  36. Milton, M.A.A. Automated Skin Lesion Classification Using Ensemble of Deep Neural Networks in ISIC 2018: Skin Lesion Analysis Towards Melanoma Detection Challenge. arXiv 2019, arXiv:1901.10802. [Google Scholar]
  37. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  38. Goyal, M.; Oakley, A.; Bansal, P.; Dancey, D.; Yap, M.H. Skin Lesion Segmentation in Dermoscopic Images With Ensemble Deep Learning Methods. IEEE Access 2020, 8, 4171–4181. [Google Scholar] [CrossRef]
  39. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  40. Adate, A.; Tripathy, B.K. A Survey on Deep Learning Methodologies of Recent Applications. In Deep Learning in Data Analytics: Recent Techniques, Practices and Applications; Acharjya, D.P., Mitra, A., Zaman, N., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 145–170. [Google Scholar] [CrossRef]
  41. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  42. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  43. Szegedy, C.; Wei, L.; Yangqing, J.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  44. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-ResNet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  45. Debgupta, R.; Chaudhuri, B.B.; Tripathy, B.K. A Wide ResNet-Based Approach for Age and Gender Estimation in Face Images. In Proceedings of the International Conference on Innovative Computing and Communications, Singapore, 24 October 2020; pp. 517–530. [Google Scholar]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  47. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
  48. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  49. Fabiano, N.; Radenovic, S. Geometric Brownian motion and a new approach to the spread of Covid-19 in Italy. Gulf J. Math. 2021, 10, 25–30. [Google Scholar]
Figure 1. Seven different types of diseases caused from lesions.
Figure 1. Seven different types of diseases caused from lesions.
Sensors 21 08142 g001
Figure 2. Occurrence of images of each type of skin cancer.
Figure 2. Occurrence of images of each type of skin cancer.
Sensors 21 08142 g002
Figure 3. Process flow diagram of the proposed method.
Figure 3. Process flow diagram of the proposed method.
Sensors 21 08142 g003
Figure 4. Confusion matrix results of different nets: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Figure 4. Confusion matrix results of different nets: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Sensors 21 08142 g004
Figure 5. Fraction classified incorrectly for all seven models: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Figure 5. Fraction classified incorrectly for all seven models: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Sensors 21 08142 g005
Figure 6. Training and validation accuracies and training and validation loss: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Figure 6. Training and validation accuracies and training and validation loss: (a) VGG19; (b) InceptionV3; (c) InceptionResNetV2; (d) ResNet50; (e) Xception; (f) MobileNet.
Sensors 21 08142 g006aSensors 21 08142 g006b
Table 1. Data augmentation of the dataset.
Table 1. Data augmentation of the dataset.
DiseaseFrequency before AugmentationMultiply Factor (k)Frequency after Augmentation
Melanocytic Nevi317913179
Benign Keratosis317103170
Melanoma165193135
Basal Cell Carcinoma126253150
Actinic Keratosis109293161
Vascular Skin Lesions46693174
Dermatofibroma281103080
Table 2. Performance of transfer learning nets without repetition of images.
Table 2. Performance of transfer learning nets without repetition of images.
Model without RepetitionAccuracyAvg. RecallAvg. PrecisionAvg. F-Measure
VGG190.67180.670.780.71
InceptionV30.81680.820.750.78
InceptionResnetV20.81140.810.820.80
ResNet500.81050.810.750.77
Xception0.80960.810.780.78
MobileNet0.82410.820.840.80
Table 3. Performance of transfer learning nets with repetition of images.
Table 3. Performance of transfer learning nets with repetition of images.
Model with RepetitionAccuracyAvg. RecallAvg. PrecisionAvg. F-Measure
VGG190.660.660.860.72
InceptionV30.790.790.870.82
InceptionResnetV20.850.860.880.86
ResNet500.770.780.860.80
Xception0.900.900.900.90
MobileNet0.880.890.880.88
Table 4. Precision, recall, F-Measure, and accuracy values of the models.
Table 4. Precision, recall, F-Measure, and accuracy values of the models.
ModelAccuracyAvg. RecallAvg. PrecisionAvg. F-Measure
VGG190.67540.67340.85480.7479
InceptionV30.86400.86190.87690.8713
InceptionResnetV20.88400.87620.87930.8845
ResNet500.82320.82220.86800.8416
Xception0.89660.89570.88760.8902
MobileNet0.87210.87110.88380.8740
Table 5. Xcpetion Net precision, recall, and F1-Score values.
Table 5. Xcpetion Net precision, recall, and F1-Score values.
DiseaseAvg. PrecisionAvg. RecallAvg. F-Measure
Melanocytic Nevi0.940.980.96
Benign Keratosis0.680.680.68
Melanoma0.580.480.52
Basal Cell Carcinoma0.880.800.84
Actinic Keratosis0.920.370.52
Vascular Skin Lesions1.00.690.82
Dermatofibroma0.710.620.67
Table 6. Test accuracy and loss values of all learning nets used.
Table 6. Test accuracy and loss values of all learning nets used.
Transfer Learning NetsAccuracyLoss
VGG1966.361.0134
Resnet5077.600.6855
InceptionResNetV285.580.6745
InceptionV379.230.6665
Xception90.480.5168
MobileNet88.570.6347
Table 7. Hardware specification.
Table 7. Hardware specification.
Hardware UseSpecification
NVIDIA GPUTesla P100
CUDA Version9.2
GPU RAM (GB)17.1
CPU ChipIntel Xeon CPU
Chip Speed (GHz)2.2 or 2.3
CPU Cores2
CPU RAM (Total GB)16.4
L3 Cache (MB)46
Disk Space (Total GB)220
Table 8. Computation time.
Table 8. Computation time.
Model NameComputational Time
(In Seconds)
VGG19746.84069
InceptionV3751.12284
InceptionResnetV22456.34356
ResNet50761.63929
Xception834.66028
MobileNet695.36065
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jain, S.; Singhania, U.; Tripathy, B.; Nasr, E.A.; Aboudaif, M.K.; Kamrani, A.K. Deep Learning-Based Transfer Learning for Classification of Skin Cancer. Sensors 2021, 21, 8142. https://0-doi-org.brum.beds.ac.uk/10.3390/s21238142

AMA Style

Jain S, Singhania U, Tripathy B, Nasr EA, Aboudaif MK, Kamrani AK. Deep Learning-Based Transfer Learning for Classification of Skin Cancer. Sensors. 2021; 21(23):8142. https://0-doi-org.brum.beds.ac.uk/10.3390/s21238142

Chicago/Turabian Style

Jain, Satin, Udit Singhania, Balakrushna Tripathy, Emad Abouel Nasr, Mohamed K. Aboudaif, and Ali K. Kamrani. 2021. "Deep Learning-Based Transfer Learning for Classification of Skin Cancer" Sensors 21, no. 23: 8142. https://0-doi-org.brum.beds.ac.uk/10.3390/s21238142

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop