Professional Documents
Culture Documents
Classification of Melanoma and Nevus in Digital Im
Classification of Melanoma and Nevus in Digital Im
Classification of Melanoma and Nevus in Digital Im
fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000.
Digital Object Identifier 10.1109/ACCESS.2017.Doi Number
ABSTRACT Melanoma is considered a fatal type of skin cancer. However, it is sometimes hard to distinguish it from Nevus
due to their identical visual appearance and symptoms. The mortality rate because of this disease is higher than all other skin
related consolidated malignancies. The number of cases is growing amongst young people but if it is diagnosed at its earlier
stage then the survival rates become very high. The cost and time required for the doctors to diagnose all patients for Melanoma
are very high. In this research work, we propose an intelligent system to detect and distinguish Melanoma from Nevus by using
state of the art image processing techniques. At first, Gaussian Filter is used for removing noise from the skin lesion of the
acquired images followed by the use of improved K-mean clustering to segment out the lesion. A distinctive hybrid super
feature vector is formed by the extraction of textural and color features from the lesion. Support Vector Machine (SVM) is
utilized for the classification of skin cancer into melanoma and nevus. Our aim is to test the effectiveness of the proposed
segmentation technique, extract the most suitable features and compare the classification results with the other techniques
present in the literature. The proposed methodology is tested on DERMIS dataset having a total number of 397 skin cancer
images where 146 are melanoma and 251 are nevus skin lesions. Our proposed methodology archives encouraging results
having 96% accuracy.
INDEX TERMS Melanoma, Nevus, Feature, K-means clustering, and centroid selection.
I. INTRODUCTION
Skin cancer is measured as a major contributor to the causes Detecting skin cancer at the initial stage can help in reducing
of deaths around the world [1]. There are various types of the risk factor in patients. Different skin cancer types can be
cancers that are discovered and battled with. However, skin found in figure 1.
cancer is amongst fast-growing cancer nowadays. According to the research, the mortality rate may be reduced
According to modern research, patients with a skin cancer up to 90 %, if the skin cancer is diagnosed at an initial stage,
diagnosis is significantly increasing more than any other Hence the diagnosis and classification of the skin cancer in
cancer form every year [2]. Melanoma is the most common its early stage are vitally important [4][5]. Among the
form of skin disease that affects the skin surface cells known conventional approaches followed by researchers to detect
as melanocytes. It consists of cells that cause the skin to turn melanoma and nevus is ABCD rule [5]. A total dermoscopic
to black color [3]. Melanoma can be found in dark or darker score is obtained for each of the ABCD features where A
color yet at some point it might likewise be in the skin, pink, represents Asymmetry, B is for Border irregularity, C
red, purple, blue or white color [4]. This form of cancer is represents color variations and D is for the Diameter. Each
very disturbing due to its tendency to cause metastasis, i.e. respective feature is assigned an individual weight based on
ability to spread. Melanoma can be found anywhere on the their significance in the feature space. Based on the
human body, however, it is mostly developed on the back of calculated score the lesion is identified as cancerous or
human legs [5]. benign. A 7-point Checklist is another technique that is used
to identify skin cancer in dermoscopic images. The list target
symptoms of atypical pigment network, grey-blue areas and
atypical vascular pattern, streaks, blotches, irregular dots and
globules, and regression patterns. At times when these
symptoms are identified, a medical professional is consulted
for the treatment [8]. Later on, the checklist reduced to a
Figure 1. Types of skin cancer lesser number of features of the different network,
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
Asymmetry and blue-white structures[9]. Considering the the characteristics of ABCD attributes and combines with
complex nature of melanoma, it becomes hard for the color asymmetry and dermoscopic structures. The proposed
researchers to detect skin cancer only on the basis of these system achieved 91.25% sensitivity. In another study[25] the
geometrical features. Another problem is that the size of the author tried to investigate the possibility to automatically
image database is increasing dramatically. So the practicality detect the dermoscopic patterns based on ABCD rule using
of such information is dependent on how well it can be deep convolutional neural networks. Experimental results
accessed, searched and how well the relevant knowledge can demonstrated 88% accurate classification of skin cancer.
be extracted from it. In another research study, Rebecca et. al [26] used ABD rule
With the advent of computer-aided diagnostic systems, excluding the color from the traditional ABCD, because
researcher mainly emphasizes on the automatic detection and color requires an additional large amount of computer
classification of skin cancer. Medical images in the form of resources. They utilized KNN as classifier achieving 89%
textural features [10]–[12], geometric features [4], [13]–[15], accuracy.
color features and in a combination [16]–[18] have been used It is also observed that ABCD rubric is sometimes very
to identify and classify skin cancer diseases. However, it is qualitative and subject manner that results in large inter-
still a challenging task to identify most discriminative observer and intra-observer bias [27]. For this purpose,
features for identifying melanoma at its initial stage [19]. High-quality intuitive features (HLIF) approach is used that
Our research work aims to achieve high accuracy results in represents the asymmetry characteristic of ABCD rubric
identifying and classifying skin cancer, contributing to the from skin cancer images [28]. The system achieved 86%
present literature by, accuracy from the extracts set of features. Later Emerald et.
al. [11] extended his previous work by proposing six new
• To develop a complete automated computer-aided HLIF features for different color channels. Experimental
system to detect melanoma cancer accurately. results show a better classification accuracy of 94%.
• Design of an improved K-Mean approach for
B. Color Features
computationally efficient segmentation.
• Utilization of hybrid features incorporating both The color features can be extracted on the basis of the
texture and color of the lesion. statistical value calculated from color channels. They are
Mean color, variance color and standard deviation value of
The remaining of our research paper is organized as a detail the RGB or HSI color model. Some of the different color
literature review of existing techniques of features extracting features extraction techniques consist of color asymmetry,
and classification is discussed in Section 2. The components centroid distance and LUV histogram distance [29][30]. In
of the proposed work are discussed in detail in section 3. another study [31], the author classified the melanoma on the
Section 4 describes the experimental setup and evaluation basis of global and local descriptors. They combined the
metrics. Section 5 portrays the results and discussion on the textural features with the color obtained classification scores
given datasets. The conclusion is given in the last section. of SE and SP are 93%, 95% respectively. Other researchers
used the same approach using color features with a set of
II. RELATED LITERATURE textural and shape-based features. Ganster et al. [6] used
color and shape based features from skin lesion with KNN
Extensive research has been done in the respective domain classifier. The number of images taken for the
by the discovery and development of new approaches to experimentation purpose is more than 5300 and they
accurately diagnose skin cancer. Related work can be achieved 87% and 92% sensitivity and specificity
divided into three types based on features extraction respectively. Rubegni et al. [32] also utilized textural and
technique i.e. geometrical, textural, and color. These geometrical features and achieved a sensitivity of 96% and
techniques are individually discussed in detail below. specificity of 96%. Celebi et al. [29] used a large features
vector consisted of color, shape and texture features. The
A. Geometrical features SVM achieved a sensitivity of 93% and specificity of 92%.
Almansour et. al in [16] used color moments with textural
The ABCD-E system [13], [15], [20], 7-point checklist [21], features with SVM as classifier achieved 90 % accuracy with
[22], 3-point checklist [23] offers geometrical features 227 melanoma and nonmelanoma skin cancer images.
techniques that are used for classification of the melanoma. .
According to Johr [14], the features extracted by the ABCD C. Textural Features
rule is computationally less expensive with respect to the 7-
point checklist. It is also observed that the consistency in the Image texture represents the spatial distribution of pixel
clinical diagnosis is at a higher rate for the ABCD rule. So, intensity levels in an image. Textural features represent the
most of the computer-aided systems for melanoma detection underlying pattern and layout of intensity levels acting as one
used ABCD rule for features extraction. Though, ABCD of the most distinctive features for object or region of interest
technique is more prone to over classification of nevi as identification. When it comes to skin cancer, textural
melanomas [24]. In research work Reda et. al.[4], extracted
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
Input Image
Classification Map
Nevus Melanoma
Classification
Pre-processing
Features Extraction
Image Segmentation
features are frequently used for image analysis as it helps in system was tested on 50 images of DERMIS dataset
classifying between nevus and melanoma by calculating the achieving 80% accuracy. Another novel approach to the
irregularity of their structure [33]. It is observed that interest Computer Aided Diagnosis (CAD) of initial melanoma has
is increasing in the computerized examination of digital appeared in the process of the web and smartphone
images taken by the Dermoscopic process. To improve early applications. In these systems, images were captured with
identification melanoma classification strategy for the high-resolution cameras instead of using an image database
computerized analysis of images obtained from ELM [38]. The system used digital camera images with context
(Epiluminescence microscopy) has been established [6]. knowledge such as type and nature of the skin, sex, age and
Region of interest was extracted using the segmentation the body part that is affected. The proposed system also tried
algorithm and a combined features strategy was used to extract the features compatible with the Dermoscopy
belonging to shape and radiometric features. KNN classifier ABCD rule. The features were further classified in various
was used to achieve 87% and 92 % sensitivity and specificity steps, a processing step for correlation-based feature
respectively. selection. The system produced specificity of 68% and a
The automatic data analysis for the melanoma early sensitivity of 94%, on images of 45 Nevus and 107
detection system (ADAM) [34] obtained 80% for both Melanoma skin cancer images. In another work [39], the
specificity and sensitivity using asymmetry and boundary author used some non-dermoscopic images of skin cancer
descriptors with (SVM) as a classifier. Iyatomi et. al. [35] where they extracted the ROI or lesion areas using k-mean
used a similar approach and achieved 100 % specificity and clustering technique and then color and textural features
95.9 % sensitivity. Recent systems in the literature [36] were extracted. Further, a set of visual features were
achieved 91 % specificity and 88.2 % sensitivity over a identified by inspecting the dermatologist. For automatic
database of 120 skin cancer images. For a proper treatment prediction the physician attributes and extraction features
of the skin cancer computer-aided diagnosis (CAD) were used. A majority vote of all the given predictions was
techniques helps the physicians to obtain a second opinion used to achieve final classification. The suggested method
about the lesion. Extracting accurate region of interest from resulted in higher diagnostic of 81% and achieved
skin lesion alongside adjacent portion is required for proper comparable results to the latest methods that are using skin
diagnosis and analysis. The authors [37] used active contour cancer images. In [40], the melanoma lesion is depicted
and watershed mask for segmentation. They extracted through a component vector containing texture data, shape,
features related to shape, textural and color. The proposed color and also global and local parameters. Some other
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
approach was presented for feature selection criterion which Where σ is the smoothing factor referring to the amount of
was built on the arrangement of differential evolution and blur.
SVM. Currently, Convolutional Neural Network (CNN) and
deep learning-based approaches were used for cancer B. SEGMENTATION
detection. The computation cost for these approaches
became barriers in clinical applications [2], [44], [45]. In the K-mean clustering is a common machine learning technique
latest research work ulzii et. al. [46] classified different skin that is extensively used in many applications such as data
cancer diseases using deep convolution network. For mining, image processing, and pattern recognition. K-mean
extracting features a Pre-trained AlexNet convolutional is considered as one of the basic methodologies for grouping
neural network model is used. An ECOC SVM classifier is and clustering of data-points into K number of clusters [18].
utilized in the classification of skin cancer. The proposed It works by splitting the image into non-overlapping groups
approach obtained the results of 90% accuracy for melanoma of pixels based on their intensity levels. The process initiates
detection. An enhanced computer automated system was by selecting the centroids from the data-points either
proposed with pixel based technique is used for randomly or through a certain criterion. The pixels or data-
segmentation [47]. They extracted features using CNN and points are clustered based on their minimum distance from
classify the images through SVM classifier achieving 93% the selected centroids. After each iteration, the mean values
accuracy on DermIS dataset. of the formed clusters are found and are set as the centroids
The afore-mentioned techniques provide promising results for the next iteration.
however, they are tested over an insufficient number of
images for detailed analysis.
From the literature, it is observed that different features
extraction techniques have been used to classify the skin
lesion. However, the combination of all these features is still
not properly tested. As the skin cancer images can be
differentiated based on color variations, therefore, color
features should be combined with other features.
III. PROPOSED METHODOLOGY Figure 3. Input image on the left with its equivalent
segmented image on the right
A. PRE-PROCESSING
Where K is the number of clusters, m is the maximum value
of a pixel in the image, and k goes from 1 to K.
Medical images are often susceptible to noise mainly due to The centroid selection technique works by ensuring
bad illumination, hair and air bubbles [48]. This inclusion of significant difference among the values of initialized
noise in images results in the formation of artifacts. Due to centroids making it more efficient and robust by converging
such artifacts, the segmentation results may get affected to the final position in a lesser number of iterations.
causing inaccurate detection results. Therefore, noise In our proposed system, input images consist of affected
removal is a significant step before applying any lesion surrounded by the background skin. The value of K is
segmentation or feature extraction technique for an accurate taken as 2, such that the foreground lesion is extracted from
diagnosis. To smoothen the image, Gaussian filter is highly the background skin as shown in Figure 3.
recommended as it removes the speckle noise added during
C. FEATURES EXTRACTION
the process of acquisition. Gaussian kernel coefficients are
sampled from the 2D Gaussian function as shown in equation
Once the lesion is segmented out of the background skin, it
1.
is then classified as malignant or benign. For better
classification results, it is required to use the best feature
descriptors for machine learning modeling. The Increase in
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
Algorithm: Skin Cancer Classification; Take skin images, perform pre-processing, segmentation and
classification
Input: RGB images; kernel; K, number of clusters; groundlabels
Output: Classification accuracy; predicted labels of skin lesions
1: procedure GuassianFiltering(img, kernel)
2: Imsmooth = gaussian (img, kernel1)
3: return Imsmooth
4: end procedure
5: procedure K-Means(Imsmooth, K)
6: (c1, c2, c3…… ck) = centroidselection(Imsmooth,K)
do
(cluster1, cluster2, ……. clusterk) = clustering((c1,c2,c3….ck),Imsmooth)
previouscentroids = (c1, c2, c3…… ck)
(c1, c2, c3…… ck) = mean(cluster1, cluster2, ……. clusterk)
newcentroids = (c1, c2, c3…… ck)
7: While newcentroids not equal to previouscentroids
9: return (cluster1, cluster2……..clusterk)
10: end procedure
the number of features increases the computational cost, with which the pixel values i and j have occurred overall
inspiring the description of precise decision boundaries. directions specified by angle θ as shown in figure 4.
Thus, it is ensured that a distinctive feature set is used.
A lesion is characterized by its texture and its color. In this
research work three different features using Local Binary
Pattern (LBP), Grey Level Co-occurrence Matrix (GLCM)
and RGB color channel features, are extracted from the ROI
of skin lesion. The techniques are utilized to extract the
textural and color-based features from the input skin lesion.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
neighbors as 1 and Euclidean distance is used. The main Four possible outcomes occur for binary class classification
disadvantage of the KNN algorithm is its complexity as it i.e. true positive, true negative, false positive and false
consumes all resources for computing the values of the negative.
features.
𝑃(𝑥|𝑐)𝑃(𝑐) ( 13)
𝑃(𝑐|𝑥) = A true positive represents when melanoma is classified by
𝑃(𝑥) the system as a melanoma while true negative happens when
Where a lesion belongs to nevus and classified as a nevus. On the
other way, a predicted false positive value happens when a
• The posterior probability of given class is P(c/x) lesion is not melanoma and it is classified as melanoma while
• The prior probability of class is P(c). a false negative happens when a certain lesion is melanoma
• The probability of predictor given class P(x|c). and is classified as nevus shown in figure 7.
• The predictor prior probability is P(x). The performance of binary class classification can be
Naïve Bayesian classifies the input hybrid feature vector x calculated in terms of its sensitivity, specificity and accuracy
into melanoma or nevus with the assumption that the quantifying its performance to false positive (FP) and false
individual attributes are independent to one another. The negative (FN) instances.
input vector is classified based on the higher posterior
probability for the two classes i.e. melanoma and nevus. • Sensitivity: The rate of positives values that are
efficiently recognized by the classifier.
iv. Decision Trees
𝑇𝑃 (14)
A decision tree is the flowchart representation of attributes 𝑆𝑒𝑛 =
in the form of a tree where each attribute acts as a node. The 𝑇𝑃 + 𝐹𝑁
tree is formed with the best attribute as the root node and then • Specificity: The ratio of negatives that are correctly
passes through each attribute following the way down to the identified by the classifier
leaf node i.e. its respective class. This results in the 𝑇𝑁 (15)
formulation of decision rules against which the test samples 𝑆𝑝𝑒𝑐 =
𝑇𝑁 + 𝐹𝑃
or records are classified. In decision trees, the instances are
classified by sorting them using a top-down approach from • Accuracy: The proportion of the number of
root to a leaf node. First, we take the root node of the tree correctly identified cases to the total number of test
and classify each instance. Then test each attribute of this cases as,
node and moving down to remaining of the tree branch 𝑇𝑃 + 𝑇𝑁
corresponding to the attribute value. The same process is 𝐴𝐶𝐶 =
(𝑇𝑃 + 𝑇𝑁 + 𝐹𝑃 + 𝐹𝑁) (16)
repeated for the sub-tree rooted at the new node.
• Precision: The proportion of predicted cases that are
IV. Datasets and Evaluation Metrics
related to malignancy.
𝑇𝑃 (17)
We have used the DermIS image database [16] in our 𝑃𝑟𝑒𝑐 =
research work. The database consists of 146 malignant 𝑇𝑃 + 𝐹𝑃
Melanoma and 251 Nevus images. The results obtained are
briefly discussed in the form of classification performance.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
100
Accuracy (%)
80
60
40
20
0
1 51 101 151 201 251 301 351
Image Number
Figure 9 Image-wise segmentation accuracy comparison using standard K-Mean and improved K-Mean
The segmented region in the skin cancer images describes 94% of the images were accurately segmented. The
cancer affected lesion. This cancerous lesion needs to be shortcoming in detection is mainly due to mismatch of
extracted from the background skin information. K-Means coordinates at borders due to low contrast between the skin
clustering using centroid selection has been utilized for the color and cancerous region. The smooth transition of
segmentation of the input images based on the variation in Melanoma affected region into the skin blurs the transition
the intensities with the value of K is set as 2. between the two intensities leading to misleading detection
The clusters formed over the images are then used for further results.
processing to accurately detect the melanoma cancer. Figure
8 shows the segmentation results using the improved K- A. Comparative Analysis with simple K-Mean
Mean technique. Ground truth data is marked on to the
images within our dataset with the help of physician and skin The k-mean clustering algorithm is utilized to extract the
cancer affected region and boundaries are identified. The K- region of interest against which ground truth labels are
Mean segmentation technique extracts the ROIs from the specified. It is of high interest that all the ground truth
images and segments out cancer affected region from the components are extracted in the process of segmentation. A
background. To measure the segmentation accuracy of a comparative analysis has been drawn between the
system, the percentage of overlap between the segmented segmentation accuracy of the proposed improved K-mean
area and the ground truth information is calculated. It is technique and the standard K-Mean. Both segmentation
achieved through the Jaccard Index. For an image with techniques have been utilized over the images within our
segmented region A, Jaccard Index of similarity with the dataset and segmentation accuracy has been recorded. It has
ground truth image B is calculated as, been observed that the K-Mean technique using centroid
A B selection proved to segment the information more
J ( A, B) = accurately. Image-wise segmentation accuracy achieved
A B over the two versions of K-Mean is shown in figure 9. It has
The total segmentation accuracy of the proposed work is been observed that for images with low the contrast between
0.94 in comparison to the utilization of standard K-Mean the foreground and background information segmentation
giving an overall index as 0.86. Using improved K-Mean results of proposed K-Mean seem to outperform the standard
one.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
CT Graph
CT of Standard K-Mean CT of Improved K-Mean
80
70
Commulative CT (sec)
60
50
40
30
20
10
0
1 51 101 151 201 251 301 351
Number of Images
Figure 10 Image wise Computational Time Analysis of standard K-Mean and Improved K-Mean
SVM KNN NB DT
Accuracy % 58 62 60 53
Sensitivity % 59 68 65 62
GLCM Features
Specificity % 50 55 52 44
Precision % 99 70 72 55
Accuracy % 69 61 70 65
Accuracy % 77 64 72 66
Accuracy % 96 84 76 94
Out of the 100 % detection of images in our dataset, proposed (CT) of segmentation utilizing standard K-Mean and
improved K-Mean showed segmentation accuracy of 94% as improved K-Mean as shown in figure 10. It has been
compared to the standard algorithm performing at 86 %. observed that the improved K-Mean segmented images in a
Standard K-Mean technique works on the random selection fewer number of iterations resulting in a low average
of intensity levels as centroids for clustering of data points. computational time of 166 milliseconds in comparison to
Initial centroid selection plays a key role in the performance 196 milli-seconds for standard K-Mean utilization.
of the algorithm. Selection of centroids through a proper B. Comparative Analysis with Separate and
mechanism ensures the initialization of centroids with Hybrid Feature
significant intensity variation. It not only improves the
segmentation results of the input data but also expedites the Multiple features are utilized to test our proposed system for
process by completing it in a fewer number of iterations. The accurate classification of skin cancer and the respective
lesser number of iterations results in an overall low results have been recorded and given in Table II. The table
computation cost of our proposed system. A comparative lists the classification accuracy of our system on individual
analysis has been drawn between the computational time features along with the classification results of the hybrid
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
feature vector. The accuracy of our system over the The binary-class classification has been of significant
classification of listed features depicts the significance of interest in the research community. Many different
those features for accurate classification. Results show that methodologies and techniques have been proposed to
the individual features distinctive they may be, however, accurately classify a lesion into cancerous or non-cancerous.
showed no significant contribution inaccuracy of the system. A comparative analysis has been made in Table III with only
Whereas, their combination showed a significant those existing methods that have utilized the same dataset as
improvement in the classification results. It also clearly ours (DermIS) providing a benchmark for our proposed
shows that the hybridization of features tends to create a
methodology. Robert Amelard [11], utilized the
more distinctive decision boundary between the two classes.
conventional approach of ABCD for classifying the lesion as
Color is an important feature that a human perceives while melanoma and nevus. They used images DermIS and
viewing or examining an image. Therefore, color can be Dermquest databases and were able to identify the test
proved as a crucial feature for distinguishing the Melanoma images 94% accurately. In another study Robert Amelard
with Nevus. Hence, the utilization of color features extricates [28] with same dataset and High-Level Intuitive Features
imperative features. Statistical results showed that with color (HLIFs) with ABCD features archives 86 % accuracy.
features only, the classification of melanoma has been
Ebtihal Almansour in [53] utilized Hybrid features as a
achieved 77% accurately. On the other hand, textural
features capture pertinent data for Melanoma identification. combination of GLCM, LBP and color features to classify
Global and textural features of GLCM and LBP respectively melanoma. The technique was evaluated over the DermIS
showed low levels of accuracy however their combined dataset with an accuracy of 90% using SVM. The technique
formation results in a better distinctive textural information. gave promising results however did not utilize the whole
Our proposed method utilizes the distinctive ability of color dataset and accuracy was calculated over only 227 images.
features along with the combination of local and global Shoieb et. al. in [47] proposed a deep learning-based feature
textural features resulting in a hybrid feature vector showing
extraction through CNN network and classifying them
promising classification results.
Visual representation of local texture and statistical values of through SVM. The algorithm was tested over multiple
global textural information is shown in figure 11. datasets however gave an overall accuracy of 93.75% over
69 images acquired through DermIS dataset. M. Ali et al. in
[37] suggested a technique to identify cancer affected skin
lesions using their textural, color-based and morphological
features. The respective features were extracted from images
acquired through DermIS dataset giving an overall
classification accuracy of 80%. The proposed method was
(a)
tested over 50 images. Rebecca et al in [26] proposed ABCD
feature-based method to classify melanoma using K-Nearest
Neighbor algorithm. The technique was tested over a few
images of DermIS dataset giving an accuracy of 89%.
Considering the above methodologies found in literature and
the information used as a baseline, existing techniques left
(b) (c)
significant room to identify the most distinctive features for
Figure 11. (a) Input Segmented Image (b) Visual representation of LBP accurate detection and classification of melanoma. The also
(c) Plot of statistical features of GLCM
showed the need to test the devised algorithm over an
As observed from Table II individual classifier accuracy adequate number of images adding more confidence to the
underperforms with respect to combined features. GLCM evaluated results. It also left with an adequate number of
features achieve accuracy of 62% with KNN as a classifier. images. To address the aforementioned shortcomings, a
LBP features perform better results compared to GLCM novel technique was developed utilizing K-mean clustering
achieving 70% accuracy. using centroid selection for efficient and better
As discussed, color is important and decrementing factor segmentation. Texture and color features were extracted
amongst all type of skin cancer, color feature achieves 77%
from the skin lesion. The features showed discriminative
accuracy. From the subsequent experiments, it is observed
by combining GLCM, LBP, and color the level of accuracy significance when utilized in combination with 96%
achieved by our system is 96%. classification accuracy.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
Clin. Dermatol., vol. 20, no. 3, pp. 240–247, 2002. [27] G. Argenziano et al., “Dermoscopy of pigmented
[15] H. Kittler, M. Seltenheim, M. Dawid, H. skin lesions: Results of a consensus meeting via the
Pehamberger, K. Wolff, and M. Binder, internet,” J. Am. Acad. Dermatol., 2003.
“Morphologic changes of pigmented skin lesions: A [28] R. Amelard, A. Wong, and D. A. Clausi, “Extracting
useful extension of the ABCD rule for high-level intuitive features (HLIF) for classifying
dermatoscopy,” J. Am. Acad. Dermatol., vol. 40, no. skin lesions using standard camera images,” in
4, pp. 558–562, 1999. Proceedings of the 2012 9th Conference on
[16] E. Almansour and M. A. Jaffar, “Classification of Computer and Robot Vision, CRV 2012, 2012, pp.
Dermoscopic Skin Cancer Images Using Color and 396–403.
Hybrid Texture Features,” IJCSNS Int. J. Comput. [29] M. E. Celebi et al., “A methodological approach to
Sci. Netw. Secur., vol. 16, no. 4, pp. 135–139, 2016. the classification of dermoscopy images,” Comput.
[17] F. Xie, H. Fan, Y. Li, Z. Jiang, R. Meng, and A. C. Med. Imaging Graph., vol. 31, no. 6, pp. 362–373,
Bovik, “Melanoma Classification on Dermoscopy 2007.
Images using a Neural Network Ensemble Model,” [30] R. J. Stanley, W. V. Stoecker, and R. H. Moss, “A
vol. 0062, no. c, pp. 1–11, 2016. relative color approach to color discrimination for
[18] M. B. B. authorSiti Z. M. H. K. Z. A. R. Hashim, malignant melanoma detection in dermoscopy
“Modified K-means Combined with Artificial Bee images,” Ski. Res. Technol., vol. 13, no. 1, pp. 62–
Colony Algorithm and Differential Evolution for 72, 2007.
Color Image Segmentation,” Comput. Intell. Inf. [31] C. Barata, M. Ruela, M. Francisco, T. Mendonca,
Syst. Springer, Cham, vol. 331, pp. 221–231, 2015. and J. S. Marques, “Two systems for the detection of
[19] M. Silveira et al., “Comparison of segmentation melanomas in dermoscopy images using texture and
methods for melanoma diagnosis in dermoscopy color features,” IEEE Syst. J., vol. 8, no. 3, pp. 965–
images,” IEEE J. Sel. Top. Signal Process., vol. 3, 979, 2014.
no. 1, pp. 35–45, 2009. [32] P. Rubegni et al., “Automated diagnosis of
[20] F. Nachbar et al., “The ABCD rule of pigmented skin lesions,” Int. J. Cancer, vol. 101, no.
dermatoscopy,” J. Am. Acad. Dermatol., vol. 30, no. 6, pp. 576–580, 2002.
4, pp. 551–559, 1994. [33] R. B. Oliveira, J. P. Papa, A. S. Pereira, and J. M. R.
[21] M. KEEFE, D. C. DICK, and R. A. WAKEEL, “A S. Tavares, “Computational methods for pigmented
study of the value of the seven‐point checklist in skin lesion classification in images: review and
distinguishing benign pigmented lesions from future trends,” Neural Computing and Applications,
melanoma,” Clin. Exp. Dermatol., vol. 15, no. 3, pp. vol. 29, no. 3. pp. 613–636, 2018.
167–171, 1990. [34] I. Stanganelli et al., “Computer-aided diagnosis of
[22] M. F. HEALSMITH, J. F. BOURKE, J. E. melanocytic lesions.,” Anticancer Res., vol. 25, no.
OSBORNE, and R. A. C. GRAHAM‐BROWN, “An 6C, pp. 4577–4582, 2005.
evaluation of the revised seven‐point checklist for [35] H. Iyatomi et al., “Computer-Based Classification of
the early diagnosis of cutaneous malignant Dermoscopy Images of Melanocytic Lesions on
melanoma,” Br. J. Dermatol., vol. 130, no. 1, pp. 48– Acral Volar Skin,” J. Invest. Dermatol., vol. 128, no.
50, 1994. 8, pp. 2049–2054, 2008.
[23] H. P. Soyer et al., “Three-point checklist of [36] Q. Abbas, M. Emre Celebi, and I. Fondón,
dermoscopy: A new screening method for early “Computer-aided pattern classification system for
detection of melanoma,” Dermatology, vol. 208, no. dermoscopy images,” Ski. Res. Technol., vol. 18, no.
1, pp. 27–31, 2004. 3, pp. 278–289, 2012.
[24] A. Masood and A. A. Al-Jumaily, “Computer aided [37] M. A. Farooq, M. A. M. Azhar, and R. H. Raza,
diagnostic support system for skin cancer: A review “Automatic Lesion Detection System (ALDS) for
of techniques and algorithms,” International Journal Skin Cancer Classification Using SVM and Neural
of Biomedical Imaging, vol. 2013. 2013. Classifiers,” Proc. - 2016 IEEE 16th Int. Conf.
[25] S. Demyanov, R. Chakravorty, M. Abedini, A. Bioinforma. Bioeng. BIBE 2016, pp. 301–308, 2016.
Halpern, and R. Garnavi, “Classification of [38] J. Fernandez Alcon et al., “Automatic imaging
dermoscopy patterns using deep convolutional system with decision support for inspection of
neural networks,” in Proceedings - International pigmented skin lesions and melanoma diagnosis,”
Symposium on Biomedical Imaging, 2016, vol. IEEE J. Sel. Top. Signal Process., vol. 3, no. 1, pp.
2016–June, pp. 364–368. 14–25, 2009.
[26] R. Moussa, F. Gerges, C. Salem, R. Akiki, O. Falou, [39] I. Giotis, N. Molders, S. Land, M. Biehl, M. F.
and D. Azar, “Computer-aided detection of Jonkman, and N. Petkov, “Expert Systems with
Melanoma using geometric features,” Middle East Applications MED-NODE : A computer-assisted
Conf. Biomed. Eng. MECBME, vol. 2016–Novem, melanoma diagnosis system using non-dermoscopic
pp. 125–128, 2016. images,” Expert Syst. Appl., vol. 42, no. 19, pp.
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI
10.1109/ACCESS.2019.2926837, IEEE Access
This work is licensed under a Creative Commons Attribution 3.0 License. For more information, see http://creativecommons.org/licenses/by/3.0/.