Next Article in Journal
FASTNN: A Deep Learning Approach for Traffic Flow Prediction Considering Spatiotemporal Features
Previous Article in Journal
Rapid Identification of Beached Marine Plastics Pellets Using Laser-Induced Breakdown Spectroscopy: A Promising Tool for the Quantification of Coastal Pollution
Previous Article in Special Issue
Numerical Approach to Facial Palsy Using a Novel Registration Method with 3D Facial Landmark
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Skin Cancer Lesions Using Explainable Deep Learning

by
Muhammad Zia Ur Rehman
1,
Fawad Ahmed
2,
Suliman A. Alsuhibany
3,*,
Sajjad Shaukat Jamal
4,
Muhammad Zulfiqar Ali
5 and
Jawad Ahmad
6
1
Department of Electrical Engineering, HITEC University Taxila, Taxila 47080, Pakistan
2
Department of Cyber Security, Pakistan Navy Engineering College, National University of Sciences & Technology, Karachi 75350, Pakistan
3
Department of Computer Science, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
4
Department of Mathematics, College of Science, King Khalid University, Abha 61413, Saudi Arabia
5
James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK
6
School of Computing, Edinburgh Napier University, Edinburgh EH10 5DT, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 6915; https://doi.org/10.3390/s22186915
Submission received: 5 August 2022 / Revised: 5 September 2022 / Accepted: 7 September 2022 / Published: 13 September 2022
(This article belongs to the Special Issue Deep Learning for Healthcare: Review, Opportunities and Challenges)

Abstract

:
Skin cancer is among the most prevalent and life-threatening forms of cancer that occur worldwide. Traditional methods of skin cancer detection need an in-depth physical examination by a medical professional, which is time-consuming in some cases. Recently, computer-aided medical diagnostic systems have gained popularity due to their effectiveness and efficiency. These systems can assist dermatologists in the early detection of skin cancer, which can be lifesaving. In this paper, the pre-trained MobileNetV2 and DenseNet201 deep learning models are modified by adding additional convolution layers to effectively detect skin cancer. Specifically, for both models, the modification includes stacking three convolutional layers at the end of both the models. A thorough comparison proves that the modified models show their superiority over the original pre-trained MobileNetV2 and DenseNet201 models. The proposed method can detect both benign and malignant classes. The results indicate that the proposed Modified DenseNet201 model achieves 95.50% accuracy and state-of-the-art performance when compared with other techniques present in the literature. In addition, the sensitivity and specificity of the Modified DenseNet201 model are 93.96% and 97.03%, respectively.

1. Introduction

According to World Health Organization (WHO) statistics, skin cancer accounts for one-third of all reported cancer cases, and the prevalence rate is increasing globally [1]. Over the past decade, a prominent increase in skin cancer has been reported in the USA, Australia, and Canada. Approximately 15,000 people do not survive every year after being infected by skin cancer [2]. An American study shows that 7180 people died in 2021 of only one type of cancer and it is expected that in the year 2022, nearly 7650 people will die because of melanoma cancer [3]. Depletion of the ozone layer increases the amount of hazardous ultraviolet (UV) radiation reaching the surface of earth. The UV radiations can damage skin cells, which may lead to the cancerous growth of cells. The hazardous UV radiation can cause a wide range of adverse consequences, and one of them is the growing incidence rate of skin cancer [4]. Some other factors including smoking, usage of alcohol, different infections, viruses, and the living environment also trigger the growth of cancerous cells. There are two major types of skin tumors; some tumors are cancerous while others are non-cancerous. The malignant tumor is cancerous, and it has further types [5]. The most prevalent malignant skin lesions are squamous cell cancer (SCC), basal cell cancer (BCC), malignant melanoma, and dermatofibrosarcoma. Numerous more include malignant cutaneous adnexae, fibrous histiocytomas, Kaposi’s sarcoma, and pleomorphic sarcoma [5,6,7]. Malignant melanoma is rare skin cancer but is considered the deadliest cancer of all of them. Malignant tumors spread to other body organs through the lymphatic system or blood vessels; this spread is called metastasis [8].
Abnormal growth of melanocytic skin cells causes malignant melanoma. Exposure of skin to sunlight produces melanin, which naturally protects the skin from adverse effects of sunlight, however, if melanin is accumulated, the tumor starts to develop [9]. In many situations, total excision (surgical removal) of an infected region performed by surgeons results in health. Moreover, in several cases, it requires rehabilitation of the infected region that is performed by plastic surgeons [7]. On the other hand, the benign tumor is non-cancerous and it does not spread to other organs but has the capability to enlarge lesion patches and tumors. Different types of benign tumors include seborrheic keratoses, cherry angiomas, dermatofibromas, skin tags (acrochordon), pyrogenic granulomas, and cysts [5,10].
In order to diagnose skin cancer, a series of steps are performed by a physician. In the first stage, lesions are inspected by the naked eye. For further examination, dermoscopy is used to further examine the pattern of skin lesions. In this process, a gel is applied to the visible skin lesions and examined under a magnifying tool for better visualization [11]. For more detailed analysis, a part of the suspected region of skin is removed and sent to a lab for microscopic examination; this procedure is referred to as a biopsy. Some experts diagnose the skin lesion based on the ABCDE technique in which a number of factors are analyzed, including color, border, asymmetry, the diameter of lesion, and development of lesion over time [12]. However, the inspection solely depends on the skill of the dermatologists along with the clinical facilities. Timely detection and diagnosis of cancer, specifically skin cancer, can avoid further spread and be cured effectively. Moreover, early detection of skin cancer helps to reduce the mortality rate and expensive medical procedures [13]. Furthermore, the manual procedure for the inspection of skin cancer is time-consuming and there could be a chance of human error during the process of diagnosis.
Since the last decade, the use of computer-aided systems has been seen in the field of medicine. Such systems can be used for the detection of skin cancer. Traditionally, different skin-related features such as color, texture, shape, etc., have been used [14]. Extraction of multiple features is not only time-consuming but also a complex process. However, recent developments in the field of artificial intelligence have paved the path for feature extraction using deep learning architectures. Deep learning architectures can extract multiple features using convolutional neural networks (CNNs) [15]. CNN can extract features efficiently as compared to traditional methods of feature extraction. Recently, deep learning-based computer-aided systems have been used for the diagnosis of different diseases and have shown remarkable results. There is a huge potential for using computer-aided systems to aid medical staff in disease diagnosis in its early stages.

2. Related Work

Recently numerous schemes have been developed for the classification of skin lesions using deep learning architectures. Some of the recent techniques for the classification of skin lesions are discussed in this section.
Dorj et al. [16] used a pre-trained AlexNet model for the process of feature extraction, while SVM was used for classification. The technique for the classification of skin lesions yielded impressive results. Filho et al. [17] presented a technique for skin lesion classification using a structural co-occurrence matrix (SCM). The SCM is used to extract texture features from dermoscopic images. Experimentation was performed on the ISIC 2016 and ISIC 2017 datasets. For classification, various learning algorithms were used, and among them, the SVM gave the best results. This technique attained a specificity of 90%. Li et al. [18] proposed a novel lesion indexing network for the classification of skin lesions. The LIN is made up of a deep learning algorithm that is capable of extracting additional features as compared to a simple deep learning algorithm. The proposed scheme attained good classification results with an accuracy of 91.2%. The scheme can also segment lesions, however, the results needed significant improvement for the segmentation of lesions. Saba et al. [19] have proposed a deep learning-based approach for the recognition of skin cancer. To improve the visual quality of the used datasets, a contrast stretching technique was used. Moreover, to estimate the boundary of lesions, a CNN followed by XOR operation was used. The features are extracted using Inceptionv3 with the help of transfer learning. For testing, the PH2 and ISIC 2017 datasets were used.
Esteva et al. [20] proposed a technique for the classification of skin cancer using the inceptionV3 model. Clinical images were used for training and evaluation purposes. The results of the proposed technique were cross-examined by a 21-member certified board for the two most deadly skin cancers. Le et al. [21] presented a deep learning technique based on the ResNet50 model that uses a transfer learning approach for model training. The hyperparameters were fine-tuned to improve the performance of the model. Moreover, to avoid overfitting, instead of the average pooling layer, global average pooling layers were used. The HAM10000 dataset was used in this work. Iqbal et al. [22] presented a technique for skin lesion classification based on deep convolutional neural networks. The model used in this work consists of multiple blocks in a top to bottom formation, which provides feature information at different scales. The model comprises 68 convolutional layers, and the ISIC 2017–2019 dataset was used in this work. Srinivasa et al. [23] proposed a technique that utilizes MobileNetV2 and LSTM networks. The proposed technique is used for lesion classification, which is performed on the HAM1000 dataset. MobileNetV2 is a lightweight model that requires low computational power and is adaptable to end devices. The LSTM network is employed to preserve temporal details of features extracted from the MobileNetV2. The combination of LSTM with MobileNetV2 improved the accuracy up to 85.34%.
Shahin et al. [24] proposed a deep learning technique based on deep convolutional neural networks for automated skin cancer classification. The proposed technique can perform binary classification for skin cancers for both benign and malignant cases. Different preprocessing steps include noise removal, normalization, and data augmentation. Several deep learning models were compared in this work to attain better and more effective classification accuracy. The proposed technique attained a test accuracy of 91.93% on the renowned HAM10000 dataset. Farhat et al. [25] presented a methodology for the classification of skin cancer employing a deep learning algorithm. Two different datasets have been used, which are HAM10000 and ISIC 2018. The proposed technique comprises a number of steps, of which the major steps are feature extraction using the deep learning model and feature selection using a metaheuristic algorithm. The final classification was performed using extreme machine learning with an accuracy of 93.40% and 94.36% for HAM10000 and ISIC2018, respectively. Chaturvedi et al. [26] proposed a multi-class skin classification technique based on deep learning models. The automated deep learning-based system tends to improve classification accuracy. Deep learning models are fine-tuned to improve the accuracy of models; moreover, ensemble models have also been used for comparison. The technique attained an accuracy of 93.20% on the HAM10000 dataset.

3. Materials and Methods

A new framework for the classification of skin lesions is presented in this section. The proposed framework can differentiate cancerous and non-cancerous lesions using deep learning models. The proposed framework requires a series of steps for the efficient classification of lesions. It starts by augmenting the available dataset and subsequently follows steps to retrain the deep learning model that includes transfer learning, fine-tuning of the model along with hyperparameter tuning. The fine-tuned model is able to extract desired features for the classification of skin cancer lesions. Two different deep learning techniques are used in this work. The augmented dataset is used for the fine-tuned deep learning model according to the requirements of this work. The general workflow of the proposed framework is presented in Figure 1. Each step of the workflow is discussed in the following subsections.

3.1. Dataset and Data Preprocessing

The dataset is the primary requirement for using deep learning models for different problems. The datasets are not readily available, and moreover, clean and well-prepared datasets are rare. Deep learning models learn the patterns and features of the dataset, and based on the learned patterns and features, they can perform predictions. A clean and well-prepared dataset is the key requirement to attaining state-of-the-art performance using deep learning models. In this work, the dataset is acquired from Kaggle [27]. Kaggle is a well-known platform for the scientific community; the dataset is part of ISIC archive [28]. The dataset consists of two categories: malignant and benign. A total of 3297 images are present in this dataset. The category “benign” contains 1800 images, while “malignant” contains 1497 images.
Data preprocessing is a vital step in enhancing the quality of any dataset [29]. It consists of multiple approaches; in this work, the visual quality of the dataset is enhanced using contrast stretching. It significantly improves the quality of images where lesion spots are faded during the process of image acquisition. A few contrasted enhanced images are shown in Figure 2.
Secondly, deep learning models require a large amount of data for training [30]. To increase the training examples of the dataset, data augmentation is used. Different data augmentation techniques are present in the literature for different scenarios. Three different augmentation techniques have been used in this work, rotation, flipping, and noise addition. Rotation and flipping are scale-invariant augmentation operations. Images are rotated at 15 and 45 degrees in both clockwise and anti-clockwise directions followed by horizontal flipping [31]. The last augmentation technique used in this work is noise addition, which is commonly known as noise augmentation. This tends to inject random noise into the dataset to increase the number of samples present in the dataset. This technique not only enlarges the dataset but also reduces the generalization error during the process of training deep learning model, making the process of training robust [32]. The Gaussian noise has been added to the dataset with a variance of 0.1. A detailed description of the dataset is presented in Table 1. A few samples of the augmented dataset are shown in Figure 3.

3.2. Deep Learning Models

Deep learning is the subfield of artificial intelligence (AI), which mimics the behavior of a human brain. During the last decade, it comes up in the spotlight when it has been used in different fields for different purposes and provided superior results compared to other existing algorithms. It is called “deep learning” as it is composed of a large number of hidden layers that are usually convolutional layers [33]. These hidden convolution layers are being used for feature extraction. Deep learning models are a step toward the automation of computer-aided systems. These models have been used for various purposes including classification, segmentation, object detection, etc., in different fields, such as agriculture, medical, driverless cars, and many more.
In this work, two deep learning models have been used for the classification of skin diseases. The used models are MobileNetV2 [34] and DenseNet201 [35]. These models are trained using Transfer Learning (TL). The TL approach not only aids the model in feature learning but also improves performance while limiting computation resources [36]. Both models have undergone minor modifications in order to achieve the desired outcomes. The used models are discussed below, along with the modifications.

3.2.1. Modified MobileNetV2

One of the models used in this work is MobileNetV2, which is a well-known model for feature extraction. Being a lightweight model, it is extensively used in the research domain. The pretrained MobileNetV2 model used in this work has been previously trained on a large image dataset, the ImageNet [37]. In this work, transfer learning is used to train the pretrained MobileNetV2 model. The original MobileNetV2 model takes the image of size 224 × 224 × 3. The image is passed through a convolutional layer having 32 filters. The inverted residual block (IRB) is the predominant block of MobileNetV2, which reduces the memory requirement as compared to the usual convolutional blocks. The IRB consists of point-wise and depth-wise convolution. The depth-wise convolution is used to eliminate redundant feature; the elimination of redundant features helps the model perform better while maintaining a low computational cost. The ReLu6 has been used as an activation function throughout the network [34]. The IRB of MobileNetV2 is commonly referred to as bottleneck; there are 17 such bottlenecks present in MobileNetV2. Figure 4 depicts the architecture of the proposed modified MobileNetV2.
In this work, three 2D convolution layers are stacked at the end of the network, which improves the performance of the model significantly. The layer CONV1_1 contains 128 filters with a kernel size of 3 × 1, the layer CONV_2 also contains 128 filters having a kernel size of 3 × 1. The final convolution layer CONV_2 contains 64 filters with a kernel size of 3 × 3. The classification head is used for final classification and is composed of a GAP layer, a batch normalization layer, and two dense layers. The final dense layer is modified according to the desired requirements.

3.2.2. Modified DenseNet201

The other model used for feature extraction is the pre-trained DenseNet201, where the number 201 refers to 201 layers of the original model. The DenseNet201 model is also trained on the ImageNet dataset. Transfer learning is used to train the model for the desired task of classifying skin diseases. The DenseNet201 model consists of 4 dense blocks and 3 transition layers that act as connections between the two dense blocks. Inside the dense block, each convolution layer is connected to other convolution layers. After every dense block, the size of the feature map is increased. The transition layers act as the downsampling layers. In DenseNet201, downsampling is performed using average pooling [35]. The architecture of the proposed Modified DenseNet201 is shown in Figure 5.
Some convolution layers are also stacked up at the end of the fourth dense block. The convolutional layer, namely the CONV1_1 layer is composed of 128 filters with a kernel size of 3 × 1 and the convolution layer CONV1_2 contains 128 filters having a kernel size of 1 × 3. The purpose of decomposing 3 × 3 kernel into 3 × 1 and 1 × 3 kernel is to reduce computational power. The third convolution layer, CONV_2 is built up with 64 filters with a kernel size of 3 × 3. The extracted features are fed into the classification head which comprises of GAP, a batch normalization layer, 2 dense layers, and the Softmax classifier. Table 2 outlines the architecture of the proposed Modified DenseNet201.

3.3. Grad-CAM Visualization

In this work, Gradient-weighted class activation mapping (Grad-CAM) visualization is used to get an insight into feature learning using the deep learning model. Generally speaking, deep learning models are considered as black boxes, as they take input and give their predictive output. Grad-CAM visualization has been used recently to understand what is happening inside the deep learning model. An illustration of Grad-CAM visualization is shown in Figure 6. It is a weakly supervised localization technique.

4. Results

The results of the proposed technique are presented in this section. Section 4.1 provides brief information on the experimental setup. The results using MobileNetV2 and DenseNet201 are presented in Section 4.2 and Section 4.3, respectively. Results are analyzed and compared in Section 4.4.

4.1. Experimental Setup

The dataset used in this work is taken from Kaggle and consists of two classes. The dataset is divided such that 70% is used for training, 20% for validation and the remaining 10% for testing. Google Colab has been used for running the MobileNetV2, DenseNet201 and their proposed modified models.

4.2. Results Based on MobileNetV2

The results using the MobileNetV2 and proposed modified MobileNetV2 are shown in Table 3. The MobileNetV2 attains an accuracy of 90.54% when trained on the dataset. The sensitivity and specificity achieved were 89.93% and 91.18%, while precision and sensitivity and F1 scores were 91.32% and 90.62%, respectively. Results of the proposed Modified MobileNetV2 are also presented in Table 3, which shows that the Modified MobileNetV2 attains an accuracy of 91.86%. Other parameters, including sensitivity, specificity, precision, and F1 scores are recorded as 91.09%, 92.66%, 92.82%, and 91.95%, respectively. Moreover, the results presented in Table 3 are also validated through the confusing matrix that is presented in Figure 7.
Figure 7a presents the confusion matrix of MobileNetV2 while Figure 7b represents the results of the proposed Modified MobileNetV2 using the confusion matrix. Figure 7a shows that MobileNetV2 model is capable of detecting benign class with an accuracy of 89%, while 11% of benign class was classified as malignant. The malignant class was accurately detected with 91% accuracy. Similarly, Figure 7b shows that the proposed Modified MobileNetV2 model detects the benign class with an accuracy of 91% and 9% of benign class was misclassified as malignant. In addition, the malignant class attains a detection accuracy of 93%, while only 7% of the malignant class was detected as benign. It can be seen from Table 3 and Figure 7 that the proposed Modified MobileNetV2 showed better performance based on the evaluation parameters. The accuracy and loss plots of the proposed Modified MobileNetV2 are presented in Figure 8.

4.3. Results Based on DenseNet201

This subsection presents the results attained using the DenseNet201 and the proposed Modified DenseNet201. The detailed results are presented in Table 4, which is based on several evaluation parameters. Table 4 shows that DenseNet201 attained an accuracy of 94.09% on the used dataset. In addition, the sensitivity, specificity, precision, and F1 score were recorded as 92.16%, 96.05%, 95.96%, and 94.02%, respectively. The proposed Modified DenseNet201 shows superiority over the pre-trained DenseNet201 by attaining an accuracy of 95.50%. To further ensure the authenticity of the obtained results, several other parameters were also considered. The sensitivity and specificity obtained by the Modified DenseNet201 model are 96.96% and 97.06%, respectively. Whereas precision and F1 score were recorded as 97.02% and 95.46%, respectively. The results shown in Table 4 are also verified using the confusion matrix shown in Figure 9.
The confusion matrix of DenseNet201 is shown in Figure 9a, whereas Figure 9b shows the confusion matrix of the proposed Modified DenseNet201. Figure 9a shows that Benign disease was correctly classified with an accuracy of 92%, while only 8% of the Benign disease was misclassified as Malignant, whereas according to the confusion matrix, the malignant disease was classified with an accuracy of 96%. In this case, the malignant disease was only 4% misclassified as Benign. Figure 9b presents the confusion matrix of the proposed Modified DenseNet201. The confusion shows that Benign and Malignant diseases were accurately classified with an accuracy of 94% and 97%, respectively. Only 6% of the benign disease was misclassified as malignant. The accuracy and loss plots of the proposed Modified DenseNet201 are shown in Figure 10.
Furthermore, a visual illustration of lesion spots is detected using the Grad-CAM technique. As discussed, the Grad-CAM weakly localized the lesion spot that can certainly aid the medical staff in detection and diagnosis. Using the proposed technique, it is evident that the learned features are well trained and are able to detect and localize lesion spots based on feature information. Figure 11 shows a few sample images that visually illustrate the purpose of Grad-CAM. The samples shown below are the results of the proposed Modified DenseNet201 with an accuracy of 95.50%. Figure 11a shows the original image, while Figure 11b shows its respective Grad-CAM based localization of the lesion spot.

4.4. Analysis and Comparison

The results in this subsection are analyzed by comparing the results of the proposed technique with the original pre-trained models. Moreover, the results are also compared with other techniques present in the literature used for the classification of skin cancer.
The accuracies attained using different models used in this work are shown in Figure 12. It is observed that the proposed Modified MobileNetV2 and Modified DenseNet201 performed better in comparison to the original pre-trained MobileNetV2 and DenseNet201 models. Moreover, as illustrated in Figure 12, the proposed Modified DenseNet201 outperforms the other three models. The proposed technique is compared with other techniques present in the literature for the classification of skin cancer. Table 5 shows that the proposed technique demonstrates its superiority over other techniques by successfully attaining 95.5% accuracy.

5. Conclusions

In this paper, the pre-trained MobileNetV2 and DenseNet201 deep learning models were modified by adding additional convolution layers to effectively detect skin cancer. Specifically, for both models, the modification includes stacking three convolutional layers at the end of both models. In addition, the classification head was modified by employing a batch normalization layer, and the final classification layer was also modified according to the class of problem under consideration. Experiments indicate that the performance of both models was increased following the architectural modifications. The Modified DenseNet201 gave the highest accuracy as compared to the other three models used in this study. The proposed Modified DenseNet201 model can be used for multi-class skin cancer diagnosis with slight changes. In addition, the optimization strategies available in the literature can be utilized for improved results.

Author Contributions

Conceptualization, M.Z.U.R.; methodology, F.A. and M.Z.U.R.; software, M.Z.U.R. and F.A.; validation, M.Z.U.R., F.A. and J.A.; formal analysis, M.Z.U.R.; investigation, S.A.A. and S.S.J.; resources, M.Z.U.R.; data curation, F.A.; writing—original draft preparation, M.Z.U.R.; writing—review and editing, F.A., S.A.A. and M.Z.A.; visualization, M.Z.U.R.; supervision, F.A. and J.A.; project administration, J.A.; funding acquisition, S.A.A. and S.S.J. All authors have read and agreed to the published version of the manuscript.

Funding

The researchers would like to thank the Deanship of Scientific Research, Qassim University for funding the publication of this project.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this research is publicly available with the name “Skin Cancer: Malignant vs. Benign”, on https://www.kaggle.com/datasets/fanconic/skin-cancer-malignant-vs-benign (accessed on 20 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. AlSalman, A.S.; Alkaff, T.M.; Alzaid, T.; Binamer, Y. Nonmelanoma skin cancer in Saudi Arabia: Single center experience. Ann. Saudi Med. 2018, 38, 42–45. [Google Scholar] [CrossRef] [PubMed]
  2. Nehal, K.S.; Bichakjian, C.K. Update on keratinocyte carcinomas. N. Engl. J. Med. 2018, 379, 363–374. [Google Scholar] [CrossRef] [PubMed]
  3. American Cancer Society. Key Statistics for Melanoma Skin Cancer. 2022. Available online: https://www.cancer.org/cancer/melanoma-skin-cancer/about/key-statistics.html (accessed on 15 March 2022).
  4. Albahar, M.A. Skin lesion classification using convolutional neural network with novel regularizer. IEEE Access 2019, 7, 38306–38313. [Google Scholar] [CrossRef]
  5. Hasan, M.R.; Fatemi, M.I.; Khan, M.M.; Kaur, M.; Zaguia, A. Comparative Analysis of Skin Cancer (Benign vs. Malignant) Detection Using Convolutional Neural Networks. J. Healthc. Eng. 2021, 2021, 5895156. [Google Scholar] [CrossRef] [PubMed]
  6. Oseni, O.G.; Olaitan, P.B.; Komolafe, A.O.; Olaofe, O.O.; Akinyemi, H.A.M.; Suleiman, O.A. Malignant skin lesions in Oshogbo, Nigeria. Pan Afr. Med. J. 2015, 20, 253. [Google Scholar] [CrossRef]
  7. Fijałkowska, M.; Koziej, M.; Antoszewski, B. Detailed head localization and incidence of skin cancers. Sci. Rep. 2021, 11, 12391. [Google Scholar] [CrossRef]
  8. Patel, A. Benign vs malignant tumors. JAMA Oncol. 2020, 6, 1488. [Google Scholar] [CrossRef]
  9. Kaur, R.; Hosseini, H.G.; Sinha, R.; Lindén, M. Melanoma Classification Using a Novel Deep Convolutional Neural Network with Dermoscopic Images. Sensors 2022, 22, 1134. [Google Scholar] [CrossRef]
  10. Akamatsu, T.; Hanai, U.; Kobayashi, M.; Miyasaka, M. Pyogenic granuloma: A retrospective 10-year analysis of 82 cases. Tokai J. Exp. Clin. Med. 2015, 40, 110–114. [Google Scholar]
  11. Marie-Lise, B.; Beauchet, A.; Aegerter, P.; Saiag, P. Is dermoscopy (epiluminescence microscopy) useful for the diagnosis of melanoma?: Results of a meta-analysis using techniques adapted to the evaluation of diagnostic tests. Arch. Dermatol. 2001, 137, 1343–1350. [Google Scholar]
  12. Redha, A.; Ragb, H.K. Skin lesion segmentation and classification using deep learning and handcrafted features. arXiv 2021, arXiv:2112.10307. [Google Scholar]
  13. Tripp, M.K.; Watson, M.; Balk, S.J.; Swetter, S.M.; Gershenwald, J.E. State of the science on prevention and screening to reduce melanoma incidence and mortality: The time is now. CA Cancer J. Clin. 2016, 66, 460–480. [Google Scholar] [CrossRef]
  14. Khan, M.A.; Alqahtani, A.; Khan, A.; Alsubai, S.; Binbusayyis, A.; Iqbal, C.M.M.; Yong, H.S.; Cha, J. Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection. Appl. Sci. 2022, 12, 593. [Google Scholar] [CrossRef]
  15. Attique, K.M.; Akram, T.; Sharif, M.; Shahzad, A.; Aurangzeb, K.; Alhussein, M.; Haider, S.I.; Altamrah, A. An implementation of normal distribution based segmentation and entropy controlled features selection for skin lesion detection and classification. BMC Cancer 2018, 18, 1–20. [Google Scholar]
  16. Dorj, U.O.; Lee, K.K.; Choi, J.Y.; Lee, M. The skin cancer classification using deep convolutional neural network. Multimed. Tools Appl. 2018, 77, 9909–9924. [Google Scholar] [CrossRef]
  17. Filho, R.; Pedrosa, P.; Peixoto, S.A.; da Nóbrega, R.V.M.; Hemanth, D.J.; Medeiros, A.G.; Sangaiah, A.K.; de Albuquerque, V.H.C. Automatic histologically-closer classification of skin lesions. Comput. Med. Imaging Graph. 2018, 68, 40–54. [Google Scholar] [CrossRef]
  18. Li, Y.; Shen, L. Skin lesion analysis towards melanoma detection using deep learning network. Sensors 2018, 18, 556. [Google Scholar] [CrossRef]
  19. Saba, T.; Khan, M.A.; Rehman, A.; Marie-Sainte, S.L. Region extraction and classification of skin cancer: A heterogeneous framework of deep CNN features fusion and reduction. J. Med. Syst. 2019, 43, 289. [Google Scholar] [CrossRef]
  20. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  21. Le, D.N.T.; Le, H.X.; Ngo, L.T.; Ngo, H.T. Transfer learning with class-weighted and focal loss function for automatic skin cancer classification. arXiv 2020, arXiv:2009.05977. [Google Scholar]
  22. Iqbal, I.; Younus, M.; Walayat, K.; Kakar, M.U.; Ma, J. Automated multi-class classification of skin lesions through deep convolutional neural network with dermoscopic images. Comput. Med. Imaging Graph. 2021, 88, 101843. [Google Scholar] [CrossRef] [PubMed]
  23. Srinivasu, P.N.; SivaSai, J.G.; Ijaz, M.F.; Bhoi, A.K.; Kim, W.; Kang, J.J. Classification of skin disease using deep learning neural networks with MobileNet V2 and LSTM. Sensors 2021, 21, 2852. [Google Scholar] [CrossRef] [PubMed]
  24. Ali, M.S.; Miah, M.S.; Haque, J.; Rahman, M.M.; Islam, M.K. An enhanced technique of skin cancer classification using deep convolutional neural network with transfer learning models. Mach. Learn. Appl. 2021, 5, 100036. [Google Scholar] [CrossRef]
  25. Afza, F.; Sharif, M.; Khan, M.A.; Tariq, U.; Yong, H.S.; Cha, J. Multiclass Skin Lesion Classification Using Hybrid Deep Features Selection and Extreme Learning Machine. Sensors 2022, 22, 799. [Google Scholar] [CrossRef]
  26. Chaturvedi, S.S.; Tembhurne, J.V.; Diwan, T. A multi-class skin Cancer classification using deep convolutional neural networks. Multimed. Tools Appl. 2020, 79, 28477–28498. [Google Scholar] [CrossRef]
  27. Skin Cancer: Malignant vs. Benign. Available online: https://www.kaggle.com/datasets/fanconic/skin-cancer-malignant-vs-benign (accessed on 20 May 2022).
  28. ISIC Archive. Available online: https://www.isic-archive.com/ (accessed on 20 May 2022).
  29. Ding, S.; Li, R.; Wu, S. A novel composite forecasting framework by adaptive data preprocessing and optimized nonlinear grey Bernoulli model for new energy vehicles sales. Commun. Nonlinear Sci. Numer. Simul. 2021, 99, 105847. [Google Scholar] [CrossRef]
  30. Khan, E.; Rehman, M.Z.U.; Ahmed, F.; Khan, M.A. Classification of Diseases in Citrus Fruits using SqueezeNet. In Proceedings of the 2021 International Conference on Applied and Engineering Mathematics (ICAEM), London, UK, 30–31 August 2021; pp. 67–72. [Google Scholar]
  31. Park, C.; Kim, M.W.; Park, C.; Son, W.; Lee, S.M.; Jeong, H.S.; Kang, J.W.; Choi, M.H. Diagnostic Performance for Detecting Bone Marrow Edema of the Hip on Dual-Energy CT: Deep Learning Model vs. Musculoskeletal Physicians and Radiologists. Eur. J. Radiol. 2022, 152, 110337. [Google Scholar] [CrossRef]
  32. Yang, J.; Lu, H.; Li, C.; Hu, X.; Hu, B. Data Augmentation for Depression Detection Using Skeleton-Based Gait Information. arXiv 2022, arXiv:2201.01115. [Google Scholar] [CrossRef]
  33. Albawi, S.; Mohammed, T.A.; Al-Zawi, S. Understanding of a convolutional neural network. In Proceedings of the 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017; pp. 1–6. [Google Scholar]
  34. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  35. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  36. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef] [Green Version]
  37. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
Figure 1. The workflow of the proposed technique for skin cancer classification.
Figure 1. The workflow of the proposed technique for skin cancer classification.
Sensors 22 06915 g001
Figure 2. Illustration of contrast-enhanced image.
Figure 2. Illustration of contrast-enhanced image.
Sensors 22 06915 g002
Figure 3. Visual illustration of the augmentation techniques.
Figure 3. Visual illustration of the augmentation techniques.
Sensors 22 06915 g003
Figure 4. Architecture of the proposed modified MobileNetV2.
Figure 4. Architecture of the proposed modified MobileNetV2.
Sensors 22 06915 g004
Figure 5. Architecture of the proposed Modified DenseNet201.
Figure 5. Architecture of the proposed Modified DenseNet201.
Sensors 22 06915 g005
Figure 6. A visual illustration of localization using Grad-CAM. (a) shows the input image and (b) depicts the lesion being weakly localized in the image using Grad-CAM. The result presented in (b) demonstrates the fact that Grad-CAM is capable of localizing lesion present in the image.
Figure 6. A visual illustration of localization using Grad-CAM. (a) shows the input image and (b) depicts the lesion being weakly localized in the image using Grad-CAM. The result presented in (b) demonstrates the fact that Grad-CAM is capable of localizing lesion present in the image.
Sensors 22 06915 g006
Figure 7. Confusion matrix of the proposed technique; (a) MobileNetV2 model, (b) Modified MobileNetV2 model.
Figure 7. Confusion matrix of the proposed technique; (a) MobileNetV2 model, (b) Modified MobileNetV2 model.
Sensors 22 06915 g007
Figure 8. Training plots of Modified MobileNetV2; (a) accuracy plot, (b) loss plot.
Figure 8. Training plots of Modified MobileNetV2; (a) accuracy plot, (b) loss plot.
Sensors 22 06915 g008
Figure 9. Confusion matrix of the proposed technique. (a) DenseNet201 model, (b) Modified DenseNet201 model.
Figure 9. Confusion matrix of the proposed technique. (a) DenseNet201 model, (b) Modified DenseNet201 model.
Sensors 22 06915 g009
Figure 10. Training plots of Modified DenseNet201 (a) Accuracy plot (b) Loss plot.
Figure 10. Training plots of Modified DenseNet201 (a) Accuracy plot (b) Loss plot.
Sensors 22 06915 g010
Figure 11. An Illustration of detection of lesion spots using Modified DenseNet201 based on Grad-CAM (a) shows the input image and (b) depicts the final output of Modified DenseNet201 using Grad-CAM.
Figure 11. An Illustration of detection of lesion spots using Modified DenseNet201 based on Grad-CAM (a) shows the input image and (b) depicts the final output of Modified DenseNet201 using Grad-CAM.
Sensors 22 06915 g011
Figure 12. Performance comparison of the four models used in this work.
Figure 12. Performance comparison of the four models used in this work.
Sensors 22 06915 g012
Table 1. A detailed description of the dataset.
Table 1. A detailed description of the dataset.
CategoriesOriginal
Images
Augmented ImagesTraining
Images
Validation ImagesTesting
Images
Benign180037272609745373
Malignant149736002520720360
Total3297732751291465733
Table 2. Architecture of the proposed Modified DenseNet201.
Table 2. Architecture of the proposed Modified DenseNet201.
LayersDenseNet201
Convolution7 × 7 conv, stride 2
Pooling2 × 2 max pool, stride 2
Dense block
(1)
[ 1 × 1   c o n v 3 × 3   c o n v ] × 6
Transition layer
(1)
1 × 1 conv
3 × 3 max pool, stride 2
Dense block
(2)
[ 1 × 1   c o n v 3 × 3   c o n v ] × 12
Transition layer
(2)
1 × 1 conv
2 × 2 average pool, stride 2
Dense block
(3)
[ 1 × 1   c o n v 3 × 3   c o n v ] × 48
Transition layer
(3)
1 × 1 conv
2 × 2 average pool, stride 2
Dense block
(4)
[ 1 × 1   c o n v 3 × 3   c o n v ] × 32
CONV1_13 × 1 conv, filters 128
CONV1_21 × 3 conv, filters 128
CONV23 × 3 conv, filters 64
Classification
Layer
Global Average Pooling
Classification Head
Table 3. Classification results using MobileNetV2.
Table 3. Classification results using MobileNetV2.
Deep Learning ModelAccuracySensitivitySpecificityPrecisionF1 Score
MobileNetV290.54%89.93%91.18%91.32%90.62%
Modified MobileNetV291.86%91.09%92.66%92.82%91.95%
Table 4. Classification results using Modified DenseNet201.
Table 4. Classification results using Modified DenseNet201.
Deep Learning ModelAccuracySensitivitySpecificityPrecisionF1 Score
DenseNet20194.09%92.16%96.05%95.96%94.02%
Modified
DenseNet201
95.50%93.96%97.06%97.02%95.46%
Table 5. Comparison with state of the art techniques.
Table 5. Comparison with state of the art techniques.
ReferencesAccuracyYear
Srinivasa et al. [23]85.34%2021
Shahin et al. [24]91.93%2021
Farhat et al. [25]94.36%2022
Proposed95.50%-
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zia Ur Rehman, M.; Ahmed, F.; Alsuhibany, S.A.; Jamal, S.S.; Zulfiqar Ali, M.; Ahmad, J. Classification of Skin Cancer Lesions Using Explainable Deep Learning. Sensors 2022, 22, 6915. https://doi.org/10.3390/s22186915

AMA Style

Zia Ur Rehman M, Ahmed F, Alsuhibany SA, Jamal SS, Zulfiqar Ali M, Ahmad J. Classification of Skin Cancer Lesions Using Explainable Deep Learning. Sensors. 2022; 22(18):6915. https://doi.org/10.3390/s22186915

Chicago/Turabian Style

Zia Ur Rehman, Muhammad, Fawad Ahmed, Suliman A. Alsuhibany, Sajjad Shaukat Jamal, Muhammad Zulfiqar Ali, and Jawad Ahmad. 2022. "Classification of Skin Cancer Lesions Using Explainable Deep Learning" Sensors 22, no. 18: 6915. https://doi.org/10.3390/s22186915

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop