Next Article in Journal
Identification of Challenges and Best Practices for Including Users with Disabilities in User-Based Testing
Next Article in Special Issue
Lung Cancer Detection Model Using Deep Learning Technique
Previous Article in Journal
Impacts of Development, Dentofacial Disharmony, and Its Surgical Correction on Speech: A Narrative Review for Dental Professionals
Previous Article in Special Issue
Syntactic Pattern Recognition for the Prediction of L-Type Pseudoknots in RNA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skin Lesion Classification Using Hybrid Convolutional Neural Network with Edge, Color, and Texture Information

1
AI Software Education Institute, Soonchunhyang University, Asan 31538, Republic of Korea
2
Department of Information Communication Software Engineering, Sangji University, Wonju 26339, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(9), 5497; https://doi.org/10.3390/app13095497
Submission received: 10 March 2023 / Revised: 13 April 2023 / Accepted: 18 April 2023 / Published: 28 April 2023
(This article belongs to the Special Issue Pattern Recognition in Biomedical Informatics)

Abstract

:
Herein, a new paradigm based on deep learning was proposed that allows the extraction of fine-grained differences between skin lesions in pixel units for high accuracy classification of skin lesions. As basic feature information for a dermoscopic image of a skin region, 50 different features were extracted based on the edge, color, and texture features of the skin lesion image. For the edge features, a line-segment-type analysis algorithm was used, wherein the visual information of a dermoscopic image was precisely analyzed in terms of the units of pixels and was transformed into a structured pattern. Regarding the color features of skin lesions, the dermoscopic image was transformed into multiple color models, and the features were acquired by analyzing histograms showing information regarding the distribution of pixel intensities. Subsequently, texture features were extracted by applying the well-known Law’s texture energy measure algorithm. Feature data (50 × 256) generated via the feature extraction process above were used to classify skin lesions via a one-dimensional (1D) convolution layer-based classification model. Because the architecture of the designed model comprises parallel 1D convolution layers, fine-grained features of the dermoscopic image can be identified using different parameters. To evaluate the performance of the proposed method, datasets from the 2017 and 2018 International Skin Imaging Collaboration were used. A comparison of results yielded by well-known classification models and other models reported in the literature show the superiority of the proposed model. Additionally, the proposed method achieves an accuracy exceeding 88%.

1. Introduction

Medical AI for aiding the diagnosis and prediction of diseases has been actively investigated to provide useful information for experts and increase diagnostic accuracy [1,2,3]. Many types of medical imaging techniques exist, such as X-ray, computed tomography, and magnetic resonance imaging. Images obtained from medical imaging exhibit high resolution and are suitable for analysis requiring high accuracy and precision. High-resolution images allow minimal feature loss during image preprocessing, such as segmentation and expSansion. In the case of skin lesion images, high-resolution images are effective for diagnosis and provide useful information. In addition, for enhanced visibility, dermatologists use high-quality magnifying dermoscopic lenses and a powerful light sources for in-vivo observations of the skin surface [4,5]. Skin diseases exhibit similar features to other types of diseases; thus, an observation method as effective as using human eyes is required. This paper presents an observation method and a classification model for the classification of seven types of lesions (actinic keratosis, basal cell carcinoma, benign keratosis-like lesions, dermatofibroma, malignant melanoma, melanocytic nevi, vascular lesions) in various skin conditions.
Skin diseases are typically diagnosed during health checkups, where a biopsy is obtained for a more detailed analysis. During a health checkup, a dermatologist visually observes the skin surface using a high-quality dermoscope and performs a diagnosis of the condition via medical consultation. If an abnormal finding is identified during the health checkup, then some or all of the lesions are removed to perform a more precise and detailed examination in a laboratory. The overall process is extremely time-consuming and costly, and the diagnosis result is fault-prone as it depends on the subjective bias of the dermatologist [6]. To address the limitations of human inspection, deep learning-based classification system utilizing big data has been proposed, e.g., ImageNet [7], MSCOCO [8], and PASCAL [9], which demonstrate high performance in terms of the detection and diagnosis of skin lesions [10,11,12,13].
During the visual observation of an object through human eyes, a significant amount of information is transmitted to the brain to allow a rapid recognition of the object [14]. In this process, numerous features of the object, such as its edge, color, and texture, are analyzed; subsequently, the information is described. Furthermore, by conducting an observation closely, a more detailed analysis can be conducted on the features of the object, allowing one to identify and understand the essence of the object. The key to accomplishing such meticulous observation is to obtain numerous features from a single image using a range of feature detectors, followed by combining the extracted features to derive the desired result. Gonzale-Luna et al. [15] extracted 137 features of breast lesions based morphological categories of shape, orientation, margin, echo pattern, and posterior. Subsequently, these extracted features were applied to machine learning algorithms, i.e., AdaBoost, K-nearest neighbors (kNNs), linear discriminant analysis, multilayer perceptron (MLP), radial basis function network, random forest, and support vector machine (SVM), and the results showed a classification performance. Derdour et al. [16] proposed a system for recognizing handwritten digits using a combination of different invariant feature extraction methods and multiple classifiers; in particular, cavities, Zernike moments, Hu moments, and the histogram of gradient were used for feature extraction. This study aimed to improve the accuracy of recognizing handwritten digits in the modified National Institute of Standards and Technology database using a Tree, a kNN, an MLP, an SVMOVO, and an SVMOVA. López-Iñesta et al. [17] proposed a hybrid approach that combines both feature extraction and feature expansion to improve classification performance based on similarity and metric learning. Feature extraction and expansion were performed using training samples obtained from data transformation and a set of standard distances. The method was used for preprocessing input data into an SVM, and the performance of the proposed method was demonstrated via four different evaluation experiments using face and object recognition data.
Based on previous studies, meticulous observations performed via machine learning typically involves the use of two or more feature detectors to extract different feature information. However, these previous studies primarily adopted hybrid approaches or systems with combined or integrated feature detector models. In this case, the throughput of the detectors and machine learning models must be high; as such, sophisticated equipment and high energy are required. The features derived from the models may be difficult to analyze depending on the detector used, and the hyperparameters of all detectors must be analyzed, which is often challenging. In general, since hyperparameters affect the final results and performance, they can only be tuned optimally by performing a series of experiments and time-consuming process.
In this paper, we focused on the basic components of images through relatively simple and easy feature analysis instead of using multiple detectors or performing large-scale models and complex processing as in previous studies [7,8,9,10,11,12,13,14,15,16,17]. These components are essential features used in the diagnosis of skin lesions, and we aimed to extract them to reduce the scale and complexity of neural models. In addition, the size of skin lesion images can always be reduced to small, fixed-size data for clear representation of the processing. Shape features based on edge information allow the analysis of the visual appearance of skin lesions, color features allow the observation of skin color changes due to disease, and texture features allow the detection of surface abnormalities of the skin caused by disease. To extract these various types of feature information in this study, shape, color, and texture features were extracted using an edge detector. Subsequently, color space conversion was performed and Law’s texture energy measure (LTEM) texture detection filter [18] was used to augment the features. Consequently, 16 edge features were generated to extract the shape information of skin lesions, and nine feature vectors were generated for color features through color space (RGB, HSV, and YCrCb) [19,20,21] conversion and histogram analysis for each channel of the respective color space. In addition, by obtaining 25 different texture features using the LTEM filter, 50 visual features were generated to be used for model training. In this case, to reduce the computational load required when a single dermoscopic image was augmented into 50 features, the augmented 50 visual features were converted into structured data via line-segment-type analysis [22,23] and histogram analysis. The converted data were used as inputs of a deep learning model for skin lesion classification, and experiments were conducted on feature augmentation and classification models.
This paper introduces existing studies on skin lesion classification in Section 2, describes feature extraction and diagnosis process in Section 3, presents the performance evaluation of the proposed method in Section 4, and finally concludes the paper with Section 5.

2. Related Studies

Skin lesions show similar appearances to a congenital nevus (mole), which poses difficulties in the early diagnosis of skin diseases via visual inspection by humans. Moreover, differences among various types of skin diseases and lesions are difficult to differentiate, which necessitates a precise analysis of skin lesions [24].
(a)
Actinic keratosis (AKIEC) presents as scaling pink/red macules or papules on areas of chronic sun exposure and is a precancerous condition of the skin that may evolve into squamous cell carcinoma (SCC).
(b)
Basal cell carcinoma (BCC), a malignant neoplasm derived from non-keratinizing cells originating in the epidermis and its basal layer, is locally invasive, and its metastasis is rare. BCC and SCC are the two most common subtypes of nonmelanoma skin cancer.
(c)
Benign keratosis-like lesions (BKLs) are benign tumors that develop on the skin, characterized by excessive formation of keratinized tissue resulting in a thick, irregular epidermal layer. Although they are generally not life-threatening, accurate diagnosis is necessary as they can be mistaken for malignant tumors.
(d)
Dermatofibroma refers to a benign tumor of the skin composed of fibrocytes and fibers forming connective tissue, often occurring in conjunction with other types of skin tumors.
(e)
Malignant melanoma (MEL) is a type of skin tumor caused by the malignant transformation of melanocytes; it may be congenital or acquired. Although MEL primarily develops on the skin, it can originate from melanocytes located in the eyes or in mucosal membranes lining the anus, nose, and esophagus.
(f)
Melanocytic nevi (NV) refer to pigmented lesions of the skin and typically occur in children or adolescents, and occasionally at birth. NV comprise a cluster of melanin-producing cells known as melanocytes. As time progresses, the number of nevi may increase, or those already present may increase in size or become darker.
(g)
Vascular lesions (VASC) are lesions such as nevus flammeus, facial flushing, telangiectasia, and hemangioma, which are caused by hemoglobin oxidation and triggers severe oxidative stress in blood cells and vessels. They appear in various forms, such as facial flushing, nevus flammeus, and rosacea.
If melanocytic nevi is suspected of being malignant melanoma, excisional biopsy is essential and, in the case of BCC or SCC, biopsy is performed to plan subsequent treatment. The examination requires approximately one week and incurs a high cost. To resolve this problem, researchers are attempting to replace conventional methods of skin disease diagnosis with medical AI [25,26,27]. Figure 1 shows eight types of skin lesions.
AI-Masni et al. [28] proposed an integrated diagnostic framework that combines the skin lesion boundary segmentation stage and the multiple skin lesion classification stage. They used a full resolution convolutional network to segment skin lesion boundaries from whole dermoscopic images; additionally, they used Inception-v3, ResNet-50, Inception-ResNet-v2, and DenseNet-201 to classify segmented skin lesions. The classification performance of each classifier model in terms of accuracy was 88.05%, 89.28%, 87.74%, and 88.70%, respectively, and the final results were obtained by synthesizing the accuracy of these classifiers. Since this framework uses multiple classifiers, the initial setup for experiments incurs high costs and the training and inference processes are time-consuming.
Barata et al. [29] classified skin lesions hierarchically into melanocytic and nonmelanocytic lesions based on their origins; subsequently, they were further classified into benign or malignant lesions depending on the malignancy degree. The classification method used may be similar to that used for binary tree problems, and DenseNet-161 and ResNet-Inception were used as the architectures of the input image encoder. For images that were subjected to feature extraction through the image encoder, a long short-term memory (LSTM) network was used to perform the process of the image decoder to derive a hierarchical diagnosis. The final accuracy was 78.0%, and the classification accuracy between melanocytic and nonmelanocytic lesions was 92.2%. Similarly, two models were used in the present study as the image encoder, and an RNN-based LSTM model was used as the decoder. LSTM requires a longer training time and inference time than CNN-based models, owing to its recurrent architecture.
By contrast, the skin disease diagnosis method proposed herein uses only a single classification model, which is achieved by extracting multiple types of features from a single image. By reducing the computational load required for visual feature extraction through dimensionality reduction, computation can be performed effectively using simple equipment. Many types of features require the segmentation of the edge image, color spaces (RGB, HSV, and YCrCb), and texture features, and each information must be converted into new one-dimensional (1D) vector information through line-segment-type analysis and histograms. In this study, 50 features in 1D vector format were generated and, after conversion, a two-dimensional (2D) matrix (50 × 256) was generated. The generated 2D matrix was used as an input for a 1D convolution-based feature analysis model. The methods for detecting each type of feature are relatively simple, and a diagnosis of a skin disease was performed by combining existing well-established methods and the line-segment-type analysis algorithm proposed by the authors in a previous study [22,23]. The proposed processing method allows the problem-solving framework of skin disease diagnosis to evolve from a classification model-centered diagnostic method (used in the past) to a feature analysis-centered diagnostic method.

3. Feature Extraction from Dermoscopic Images

In this study, to extract edge, color, and texture features from a dermoscopic image for skin lesion classification, fine-grained edge detection algorithm, color space conversion, and texture detection using the LTEM filter were performed.

3.1. Extraction of Edge Feature Information via Line-Segment-Type Analysis

To perform high-precision analysis of visual characteristics in skin lesions, we used edge detection methods that can extract features from small internal areas, rather than the typical contour processing methods.
Figure 2 shows the borders of a skin lesion. The outer-border (OB) represents the border between the normal skin region and the lesion, and the inner-border (IB) inside the lesion represents the border information in the center of the lesion. Since conventional edge detectors identify edge features for strong line segments regardless of the location of the skin lesion in the outer or inner borders, small areas in the IB tend to be removed, which results in reduced accuracy in feature analysis. Therefore, in this study, when detecting the edge information of a skin lesion, the brightness was removed to differentiate the OB and IB of the lesion. In general, lesions show a weaker intensity (darker) than normal skin regions, and a threshold value for removing areas other than the lesion can be calculated using Equation (1).
q 1 = M I N I + I ( x , y ) 2 N × M I ( x , y ) N × M 2 q 2 = M A X I I ( x , y ) 2 N × M I ( x , y ) N × M 2
I x , y in Equation (1) denotes the intensity of the image at position x , y of the dermoscopic image, and N × M indicates total number of pixels, which represents the size of the image. After calculating the minimum and maximum values of the I x , y pixel of the input dermoscopic image, the lower and upper threshold values ( q 1   a n d   q 2 , respectively) are generated using the standard deviation of I x , y . The values of q 1   a n d   q 2 are set to be similar to the mean pixel value by adding the standard deviation to the minimum value and subtracting the standard deviation from the maximum value.
Figure 3a presents the skin lesion image and Figure 3b presents the histogram of Figure 3a. In Figure 3a, the normal skin appears brighter than the lesion and belongs to the uppermost section of the histogram, whereas the darker region in the lesion belongs to the lowest section of the histogram. Using the threshold range [ q 1 , q 2 ] , the uppermost and lowermost regions of the input image can be removed using Equation (2).
Based on Equation (2), if the pixel value of I x , y is within the threshold range [ q 1 , q 2 ] , then the original brightness value at the corresponding position is assigned; if the pixel value is out of the range, then it is set to 0, and the threshold image I T ( · ) is generated. Consequently, the OB and IB of the lesion are clearly revealed as the normal skin region, and the regions showing low brightness inside the lesion are removed, as shown in Figure 3c. Figure 3d presents the histogram of I T ( · ) , which is an image with the threshold range values applied. As shown, the uppermost and lowermost regions have been removed in the histogram.
I T ( x , y ) = I ( x , y ) i f   I x , y   e x s i t s   i n   [ q 1 , q 2 ] 0 o t h e r w i s e .
Otsu’s algorithm [30] was applied to I T ( · ) in Figure 3c, where some regions of brightness were removed via thresholding, and the IB and OB were extracted. Otsu’s algorithm is a method for dividing an object region in an image. Using a histogram, the algorithm allows the intensity to be calculated to achieve the most optimal classification of the image into two classes.
Algorithm 1 is a method of detecting IB and OB regions based on Otsu’s algorithm. The optimal intensity k is calculated using Otsu’s algorithm. Subsequently, α is calculated using the mean m G and standard deviation σ G 2 of the image I T ( · ) , and the optimal k and k are generated. Figure 4a shows the skin lesion image, and the image obtained after applying image thresholding using Equations (1) and (2) is shown in Figure 4b. When the k derived using Otsu’s algorithm is applied to image (b), the OB information is obtained as shown in Figure 4c. Subsequently, by applying k , the IB image is obtained, as shown in Figure 4d. Figure 4e,f show edge images detected by the Canny edge detector from the images shown in Figure 4c,d, respectively. Figure 4e shows the extraction of the IB region, and Figure 4f shows the line-segment feature of the OB region. Finally, the image shown in Figure 4g is generated by the XOR operation of the images shown in Figure 4e,f.
In Figure 4b, the normal skin region was removed via the previous processing; however, the normal skin region remained around the lesion. Therefore, Otsu’s algorithm was used to distinguish between the normal skin region and lesion such that only the normal skin region is extracted, as shown in Figure 4c. As for the regional separation, the boundary lines closer to the center represent the characteristics inside the lesion, while the boundary lines further away from the center form the outer shape of the lesion. After extracting the edges from the images shown in Figure 4c,d using the Canny detector (i.e., (c) -> (e), (d) -> (f)), an XOR operation was performed to remove the border of the normal skin region, which is the common area in the two images, and the resulting image is presented in Figure 4g.
As shown in Figure 4g, when an edge image with visual shape information of the IB and OB for the skin lesion was generated, line-segment-type analysis was performed in pixel units using a detector filter based on Equation (3).
Algorithm 1: Inner and Outer Border Division Algorithm Using Otsu’s Algorithm
Input: I T x , y Image with removed brightness via thresholding I x , y ; image size, N × M
Output: Inner and outer border-image I B ( x , y ) and edge image I E ( x , y ) of I B ( x , y )
Generate an array h i s t [ L ] of size L to create a histogram distribution diagram
Use the brightness intensity of input I T ( x , y ) as the index of array h i s t [L] and increase the value of the corresponding index by 1.
Repeat Step ② to scan all the pixels of the input I T x , y .
For array h i s t [ i ] that has completed step ③, obtain the normalized histogram value is using p i = h i s t i N × M .
Calculate the parameters required for the classification into two classes:
-
The probability that a specific pixel belongs to Class 1 P 1 k = i = 0 k p i
-
The probability that a specific pixel belongs to Class 2 P 2 ( k ) = 1 P 1 ( k )
-
The mean of the brightness of pixels in Class 1, m 1 k = 1 P 1 k i = 0 k i p i
-
The mean of the brightness of pixels in Class 2, m 2 k = 1 P 2 ( k ) i = k + 1 L 1 i p i ,
-
The mean intensity up to the intensity level k, m k = i = 0 k i p i
-
The global mean of the image, I T ( x , y ) m G = i = 0 L 1 i p i
-
The global variance of the image, I T ( x , y ) σ G 2 ( k ) = i = 0 L 1 ( i m G ) 2 p i
Using the between-class variance σ B 2 k = [ m G P 1 k m ( k ) ] 2 P 1 k [ 1 P 1 ( k ) ] , obtain the value of k for maximizing the variance.
Using the optimal value of k calculated in Step ⑥, derive k = k α using α = m G σ G 2 .
I B O ( x , y ) is the image of the outer-border division through k in I T ( x , y ) , I B I ( x , y ) is the image of inner-border division through k
Create edge image I E O x , y of I B O x , y and edge image I E I x , y of I B I x , y
Perform an XOR operation on the two edge images to generate the inner- and outer-border edge image I E x , y .
L T D F = 1 2 4 8 0 16 32 64 128 .
The line-segment-type detector filter (LTDF) expressed in Equation (3) is a detector filter, the coefficients of which are composed of values of 2 N . Furthermore, a convolution operation with the binary edge image generated for edge information extraction was performed with pixel 1 in the image. As an example, Figure 5a shows a partial region in the input image.
When convolution is performed with the LTDF shown in Figure 5b, the shaded coefficient values in Figure 5b are used in the operation. In other words, the visual line-segment information in the image can be classified into a binary code format in a structured pattern (e.g., [8, 32, 64, 128]), as shown in Figure 5c, and the filtering result can be used as a unique type number for the corresponding line segment.
In this study, the visual information converted to structured information using the LTDF was classified into 256 types, as shown in Table 1. When the structured information was classified into similar line-segment types, the classification result can be organized as presented in Table 2.
Table 1 shows the line-segment types that can be classified using the LTDF, and the numbers in Table 1 indicate the unique numbers (eigen-numbers) based on the line-segment type. For the same line-segment type, different eigen-numbers are assigned depending on the position. Meanwhile, general line-segment types, such as vertical, horizontal, and curved, as well as unusual line types such as checkered, are expressed as Pattern N. A total of 19 different patterns were generated, as shown in Figure 6.
In this study, the values calculated using the LTDF and filtering were set as the eigen-numbers to reduce the visual edge information in the image. At this time, the input dermoscopic image was segmented into regions of the same area based on Equation (4), and the LTDF was applied to each segmented region. The pixel value in the segmented image, which is the value obtained using the LTDF, corresponds to the eigen-number in Table 1.
In Equation (4), both M   a n d   N denote the size of the input image I E ( · ) , α indicates the number of vertical and horizontal segmented regions of equal area, and · denotes the floor() function. In this study, a single image is partitioned into 16 regions, and their line-segment type value is calculated using the LTDF. For each line-segment type, Equation (5) is used to generate an E F A i , j · vector, which is the edge feature for a single segmented region S i , j ( · ) .
S i , j x , y = I E S w × i + x , S h × j + y w h e r e , S w = M / α , S h = N / α , 0 x < S w , 0 y < S h
E F A i , j · = a 0 , a 1 , , a k , , a K ,   i , j = 0 , 1 , , α 1 , K = 255 w h e r e , a k = C u m k ,   k = S i , j ( x , y ) L T D F .
E F A i , j [ · ] in Equation (5) represents the reduced visual edge information, K the maximum number of line-segment types, and the filtering operation in the image processing. The function C u m ( · ) indicates a cumulative function, which returns the cumulative number of pixels whose filtering result obtained using the LTDF is k in the segmented region S i , j ( · ) . After calculating the values of E F A i , j [ · ] for all the segmented regions of the dermoscopic image, the results were converted into i × j × K vectors through normalization and used as edge information for the learning model.

3.2. Extraction of Color Feature Information from Color Space Histogram

To extract the color information, which is another visual information of skin lesions, the image was converted to RGB, YCbCr, and HSV color models, which are representative color spaces generally used in object segmentation and classification. Additionally, histogram analysis was performed on each channel of the respective converted color models for feature extraction. The YCbCr color model represents chromaticity components in terms of luminance (Y), blue-difference (Cb), and red-difference (Cr). The HSV model comprises hue (H), saturation (S), and value (V) components, which clearly distinguishes luminance and chrominance, and is typically used in color image segmentation because its color space components are the most intuitive to human recognition. Figure 7 shows an example of a dermoscopic image converted to the three color models.
When a dermoscopic image was converted to three color spaces, the chrominance of the lesion was clearly generated in contrast to normal, as shown in Figure 7, and the color information vector C F M [ · ] was generated as shown in Equation (6) using histogram information.
C F A · = C 1 H R , H G , H B , C 1 H Y , H C r , H C b , C 1 H H , H S , H V , w h e r e , H m = h 0 , , h k , , h K , m = R , G , B , Y , C r , C b , H , S , V , ( K = 255 ) h k = H i s t o g r a m ( k ) .
In Equation (6), C F A [ · ] represents the color information corresponding to three channels for each of the three color models, and it is represented as a vector with a size of 3 (color model) × 3 (number of channels) × 256 (histogram), thus resulting in nine vectors with 256 elements. h k is an element of H m [ · ] , which is a histogram information vector of one channel and indicates the value of level k in the channel’s histogram.

3.3. Extraction of Texture Features Using LTEM Filter

The LTEM algorithm is typically used to extract the text features of various objects and allows texture information from dermoscopic images to be extracted with relative simplicity. As shown in Table 3, for masks used to extract texture features from the image presented herein, five-value vectors that can extract texture information such as level, edge, spot, wave, and ripple were used to obtain the masks (as shown in Equation (7)).
M i , j = V i T × V j = a 1 a 2 a k × b 1 b 2 b k = a 1 b 1 a 1 b k a k b 1 a k b k V i = a 1 , a 2 , , a k , V j = b 1 , b 2 , , b k .
In Equation (7), V i [ · ] denotes a mask vector corresponding to label i shown in Table 3, and a 2D spatial filter for texture feature extraction was generated by the vector product of five 1D mask vectors. Here, M i , j denotes a texture feature extraction filter generated by two 1D masks corresponding to labels i and j. Therefore, the texture feature T F A [ · ] of the dermoscopic image extracted for learning is defined as shown in Equation (8) below.
T F A · = H 1,1 H 1 , J H I , 1 H I , J w h e r e ,   i f   i   a n d   j   i s   1 ,         H i , j = h 0 , , h k , , h K ,   K = 255 h k = H i s t o g r a m k   o f   T i , j x , y = I T x , y M i , j o t h e r w i s e   H i , j = E F A ·   o f   T i , j x , y = I T x , y M i , j .
In Equation (8), H i , j [ · ] denotes the value corresponding to the histogram or line-segment-type features of image T i , j · , which is obtained by extracting texture information from the dermoscopic image by filtering the I T · image through M i , j . The value of M 1,1 L 5 L 5 is a value that represents relatively high brightness. A feature vector was generated using the histogram value, whereas for the other texture features, the feature vectors were generated using the line-segment-type analysis method, which was discussed in the previous section. Figure 8 shows a filtering result image for 25 texture masks (5 × 5) used for texture feature extraction in this study.
By performing the process above, a feature vector for a skin lesion was generated with sixteen edge features ( E F A [ · ] ), nine color features ( C F A [ · ] ), and twenty-five texture features ( T F A [ · ] ) from a single dermoscopic image, as shown in Figure 9. The generated single feature data exhibit the same vector size with 256 elements. Therefore, the input pattern data of the artificial neural network for skin lesion recognition, X [ · ] , was generated as a single vector of size 50 × 256 and flattened based on Equation (9) below. Subsequently, they were used as the input of the neural network model.
X · = E F A C F A T F A = E F A 1,1 E F A 1,4 E F A 4,1 E F A 4,4 H R H G H B H Y H C b H C r H H H S H V H 1,1 H 1,5 H 5,1 H 5,5 = E F A 1,1 E F A 4,4 H R H V H 1,1 H 5,5

4. Evaluation of Model Performance

In this study, datasets from the 2017/2018 International Skin Imaging Collaboration (ISIC) were used to evaluate the performance of the proposed feature extraction method and learning model used in this study. For the proposed model, which uses structured data generated via feature extraction from a dermoscopic image as training data, the performance of the optimization function was first evaluated based on the learning rate. Subsequently, the model performance was evaluated via comparison with state-of-the-art image classification models and results pertaining to skin lesion classification reported in the literature.

4.1. Experimental Setup

As shown in Figure 9, the proposed algorithm extracts edge, color, and texture features from a single dermoscopic image, and then generates data with reduced dimensionalities smaller than 256 × 256, which is the general the data size of a raw image (as shown in Equation (9)). In the generated data, strong feature components shown in the image can be represented with emphasis, and weak feature components can be represented in terms of the number of cumulative pixel intensities, which allows a representation with an aggregated feature component.
In this study, to perform machine learning on the input feature data extracted from a dermoscopic image, as shown in Figure 10a, we established four convolution groups for the model: Conv-Ag, Conv-Bg, Conv-Cg, and Conv-Dg. As shown in Figure 10b, each convolution group was a modular form with a 1D convolution layer connected to a 1D global max-pooling layer, where N modules were not connected to each other but configured in parallel. In this case, each 1D convolution layer had a different kernel size and stride, and the ReLU function was used as an activation function. Among the convolution groups, the padding of A and C was set as “same,” and the padding of B and D was set as “valid.” For convolution groups A and B, the analysis was performed based on the features of the input data, as shown in Figure 9, whereas for convolution groups C and D, the rows and columns of the input data were transposed to analyze the differences of each feature.
Each set of data calculated from the convolution groups was concatenated and integrated into one set of data, and after performing learning for each group using fully connected network-1 (FN-1), all data were integrated via a second concatenation. In FN-2, the strong features of the second concatenated data were analyzed. Finally, the data integrated via the third concatenation were classified into the “softmax” type.
In this study, to validate and evaluate the performance of the proposed feature extraction method and learning model, datasets from ISIC 2017 [31] and 2018 [32] were used. In particular, the performance was evaluated based on the annotation information of each set of data obtained from the ISIC datasets. The ISIC 2017 dataset comprised 2000 skin lesion images from three categories: “nevus (1372)”, “melanoma (374)”, and “seborrheic keratosis (254).” Meanwhile, the ISIC 2018 dataset comprised 10,015 images from seven categories: “actinic keratosis (327)”, “basal cell carcinoma (514)”, “benign keratosis-like lesions (1099)”, “dermatofibroma (115)”, “melanoma (1113)”, “melanocytic nevi (6705)”, and “vascular lesions (142).”
As for the hardware used in the experimental setup, an Intel Core i7-6700 CPU with 32 GB of RAM and GeForce RTX 2080 Ti were used on a Windows 10 operating system. To implement the feature extraction and machine learning algorithms, Numpy1.18 and OpenCV4.1 were used based on Python3.6, Keras2.1, and Tensorflow2.4 libraries.

4.2. Experimental Results

Table 4 shows the experimental results for the optimal learning of the proposed model using the ISIC 2018 dataset. Among 10,015 images, 80% were used as the training dataset (8012 images) and 20% as the test dataset (2003 images). The result was measured using the EarlyStopping() function and the Keras-tuner library. Meanwhile, the performance was evaluated at learning rates of 0.01, 0.001, 0.0001, and 0.00001. The result shows that the AdaGrad function demonstrated the highest accuracy of 0.88 at learning rates of 0.01 and 0.001, and that the loss of the AdaGrad function was insignificant (lr = 0.00001). Figure 11 presents a confusion matrix of the AdaGrad optimization function (lr = 0.001).
Table 5 presents the result of the performance comparison between the proposed model and state-of-the-art image classification models. For each of the models compared, the size of the dermoscopic image was normalized to 512 × 512, 256 × 256, and 128 × 128. Meanwhile, for the proposed model, a 1D convolution layer-based model was used with input pattern data of size 50 × 256, generated through an analysis of edge, color, and texture features. The performance of the proposed model was compared with those of DenseNet-201, DenseNet-121, SEResNeXt-101, SEResNeXt-50, ResNet-18, ResNet-50, and ResNet-152. The result shows that the accuracy of the proposed method was 88.6%, which was the highest among all the models compared.
In this study, image processing in pixel units was performed to observe skin lesions more closely. The size of the input data generated by this method was 50 × 256, which is smaller than that of other models using data with sizes of 512 × 512, 256 × 256, and 128 × 128 pixels. In addition, despite the small input data size, the proposed method contains essential feature information regarding the edge, color, and texture features; therefore, the proposed method enables a higher-level analysis compared with the processing method using only the original dermoscopic image.
Table 6 presents a comparison between methods used in previous studies [33,34,35,36] that achieved an excellent performance in skin lesion classification using the ISIC 2018 dataset and the proposed method. As shown, the proposed method outperformed the other methods. Although the result of Minjie shows the highest accuracy at 89.5%, the proposed method is useful and convenient for feature analysis in pixel units, as the method clearly shows the differences of fine-grained features through the detection of edge, color, and texture features, thereby enhancing the classification performance.
Table 7 shows the results of performance validation using the ISIC 2017 dataset. For the existing model [37], a local binary pattern was applied to the original image to obtain the results. As shown in Table 7, the proposed method yielded a high accuracy of 89.3% in melanoma recognition, as well as demonstrating a high performance in terms of the area under the curve (AUC) (87.1%) and sensitivity (SENS) (89.0%). In terms of seborrheic keratosis recognition, the accuracy of the proposed model was low (i.e., 86.3%); similarly, its AUC (91.6%) and specificity (86.3%) were low, although its SENS was higher (86.0%) than that of the other models. Table 8 summarizes the comparison between the results yielded by the proposed method and those reported in previous studies [31,37,38,39].

5. Conclusions

For the classification of skin lesions in dermoscopic images, we herein proposed a detailed feature extraction method that analyzes edge, color, and texture features in pixel units. The proposed method applies the line-segment-type analysis algorithm to present the line-segment information of edges based on different types. Additionally, the type of visual lines was analyzed in pixel units, and line-segment features were extracted based on cumulative statistics for the same intensity. Regarding color feature extraction for detecting fine-grained changes in color for skin lesions, the image was converted into RGB, HSV, and YCbCr color spaces to analyze the histogram distribution of each channel. Furthermore, to extract texture features, an LTEM filter was used to extract 25 different elements of texture information, and the features were extracted using histogram analysis and a line-segment-type analysis algorithm. Using the input data generated through this process, machine learning was performed via a learning model designed based on a 1D convolution layer. In the designed model, 1D convolution layers with different learning parameters were connected in parallel to determine the optimal weights for the input data.
The performance of the proposed feature data generation method and learning model was compared and evaluated with other well-known classification models using the ISIC 2017 and ISIC 2018 datasets. In experiments using the ISIC 2018 dataset, the proposed method showed a superior performance with an accuracy of 94.9%, compared to previous research results. On the other hand, for the recognition of melanoma and seborrheic keratosis using the ISIC 2017 dataset, the proposed method showed a comparable performance of 97.1% with the other models. Additionally, when compared to results reported in the literature with excellent performance, the proposed method demonstrated an almost equivalent performance.
The method proposed herein can detect fine-grained variations in skin lesion features, and in situations requiring the learning of numerous features, since the proposed method uses extracted data, a reduced computational load can be expected as compared with models using an entire image as the input. However, a significant amount of feature extraction is required for detecting regions of skin lesion with fine-grained features. To effectively extract numerous features, the feature extraction process must be further simplified to increase the speed of data preprocessing and model training. In addition, the experiment revealed that there is a possibility of losing significant feature information during the process of normalizing image sizes, which may result in reduced performance. Therefore, additional experiments are needed to address this issue. The proposed feature extraction method and learning model for detecting and classifying skin lesions may serve as a novel approach for deep learning-based image recognition. Furthermore, they may yield an excellent performance in other fields that require the detection of fine-grained visual differences in input images.

Author Contributions

Investigation, M.J., Y.H. (Younghwan Han), Y.H. (Yousik Hong); writing—original draft preparation, C.K.; writing—review and editing, W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Sangji University Research Fund, 2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahmadi, A.; Kashefi, M.; Shahrokhi, H.; Nazari, M.A. Computer aided diagnosis system using deep convolutional neural networks for ADHD subtypes. Biomed. Signal Process. Control. 2021, 63, 102227. [Google Scholar] [CrossRef]
  2. Mansour, R.F. Deep-learning-based automatic computer-aided diagnosis system for diabetic retinopathy. Biomed. Eng. Lett. 2018, 8, 41–57. [Google Scholar] [CrossRef] [PubMed]
  3. Lee, H.; Chen, Y.-P.P. Image based computer aided diagnosis system for cancer detection. Expert Syst. Appl. 2015, 42, 5356–5365. [Google Scholar] [CrossRef]
  4. Yang, W.; Zhou, F.; Zhu, R.; Fukui, K.; Wang, G.; Xue, J.H. Deep learning for image super-resolution. Neurocomputing 2019, 398, 291–292. [Google Scholar] [CrossRef]
  5. Ahn, E.; Kim, J.; Bi, L.; Kumar, A.; Li, C.; Fulham, M.; Feng, D.D. Saliency-based lesion segmentation via background detection in dermoscopic images. IEEE J. Biomed. Health Inform. 2017, 21, 1685–1693. [Google Scholar] [CrossRef] [PubMed]
  6. Milton, M.A.A. Automated skin lesion classification using ensemble of deep neural networks in ISIC 2018: Skin lesion analysis towards melanoma detection challenge. arXiv 2019, arXiv:1901.10802. [Google Scholar]
  7. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  8. Vinyals, O.; Toshev, A.; Bengio, S.; Erhan, D. Show and tell: Lessons learned from the 2015 mscoco image captioning challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 652–663. [Google Scholar] [CrossRef]
  9. Shetty, S. Application of convolutional neural network for image classification on Pascal VOC challenge 2012 dataset. arXiv 2016, arXiv:1607.03785. [Google Scholar]
  10. Pollastri, F.; Parreño, M.; Maroñas, J.; Bolelli, F.; Paredes, R.; Ramos, D.; Grana, C. A Deep Analysis on High Resolution Dermoscopic Image Classification. Orig. Res. Pap. 2021, 15, 514–526. [Google Scholar] [CrossRef]
  11. Yao, P.; Shen, S.; Xu, M.; Liu, P.; Zhang, F.; Xing, J.; Xu, R.X. Single model deep learning on imbalanced small datasets for skin lesion classification. arXiv 2021, arXiv:2102.01284. [Google Scholar] [CrossRef] [PubMed]
  12. Saeed, J.; Zeebaree, S. Skin Lesion Classification Based on Deep Convolutional Neural Networks Architectures. J. Appl. Sci. Technol. Trends 2021, 2, 41–51. [Google Scholar] [CrossRef]
  13. Bhardwaj, A.; Rege, P.P. Skin lesion classification using deep learning. In Advances in Signal and Data Processing; Springer: Singapore, 2021; pp. 575–589. [Google Scholar]
  14. Wang, L.; McAlonan, K.; Goldstein, S.; Gerfen, C.R.; Krauzlis, R.J. A Causal Role for Mouse Superior Colliculus in Visual Perceptual Decision-Making. J. Neurosci. 2020, 40, 3768–3782. [Google Scholar] [CrossRef] [PubMed]
  15. González-Luna, F.A.; Hernández-López, J.; Gómez-Flores, W. A performance evaluation of machine learning techniques for breast ultrasound classification. In Proceedings of the 2019 16th International Conference on Electrical Engineering, Computing Science and Automatic Control (CCE), Mexico City, Mexico, 11–13 September 2019; pp. 1–5. [Google Scholar]
  16. Derdour, K.; Mouss, H.; Bensaadi, R. Multiple Features Extraction and Classifiers Combination Based Handwriting Digit Recognition. Int. J. Electr. Eng. Informatics 2021, 13, 163–178. [Google Scholar] [CrossRef]
  17. López-Iñesta, E.; Grimaldo, F.; Arevalillo-Herráez, M. Combining feature extraction and expansion to improve classification based similarity learning. Pattern Recognit. Lett. 2017, 93, 95–103. [Google Scholar] [CrossRef]
  18. Laws, K.I. Texture energy measures. In Proc. Image Understanding Workshop; Science Applications Inc.: Los Angeles, CA, USA, 1979; pp. 47–51. [Google Scholar]
  19. Liao, J.; Wang, Y.; Yin, J.; Liu, L.; Zhang, S.; Zhu, D. Segmentation of Rice Seedlings Using the YCrCb Color Space and an Improved Otsu Method. Agronomy 2018, 8, 269. [Google Scholar] [CrossRef]
  20. Zhang, X.-N.; Jiang, J.; Liang, Z.-H.; Liu, C.-L. Skin color enhancement based on favorite skin color in HSV color space. IEEE Trans. Consum. Electron. 2010, 56, 1789–1793. [Google Scholar] [CrossRef]
  21. Pennisi, A.; Bloisi, D.D.; Nardi, D.; Giampetruzzi, A.R.; Mondino, C.; Facchiano, A. Skin lesion image segmentation using Delaunay Triangulation for melanoma detection. Comput. Med. Imaging Graph. 2016, 52, 89–103. [Google Scholar] [CrossRef]
  22. Kim, C.-M.; Hong, E.J.; Chung, K.; Park, R.C. Driver Facial Expression Analysis Using LFA-CRNN-Based Feature Extraction for Health-Risk Decisions. Appl. Sci. 2020, 10, 2956. [Google Scholar] [CrossRef]
  23. Kim, C.M.; Lee, W.B. Line-Segment Feature Analysis Algorithm for Handwritten-Digits Data Reduction. KIPS Trans. Softw. Data Eng. 2021, 10, 125–132. [Google Scholar]
  24. ASAN Medical Center. Available online: http://eng.amc.seoul.kr/gb/lang/main.do (accessed on 26 October 2020.).
  25. Shimizu, K.; Iyatomi, H.; Celebi, M.E.; Norton, K.-A.; Tanaka, M. Four-Class Classification of Skin Lesions with Task Decomposition Strategy. IEEE Trans. Biomed. Eng. 2014, 62, 274–283. [Google Scholar] [CrossRef] [PubMed]
  26. Sumithra, R.; Suhil, M.; Guru, D. Segmentation and Classification of Skin Lesions for Disease Diagnosis. Procedia Comput. Sci. 2015, 45, 76–85. [Google Scholar] [CrossRef]
  27. Kassem, M.A.; Hosny, K.M.; Fouad, M.M. Skin Lesions Classification into Eight Classes for ISIC 2019 Using Deep Convolutional Neural Network and Transfer Learning. IEEE Access 2020, 8, 114822–114832. [Google Scholar] [CrossRef]
  28. Al-Masni, M.A.; Kim, D.H.; Kim, T.S. Multiple skin lesions diagnostics via integrated deep convolutional networks for segmentation and classification. Comput. Methods Programs Biomed. 2020, 190, 105351. [Google Scholar] [CrossRef]
  29. Barata, C.; Marques, J.S.; Emre Celebi, M. Deep attention model for the hierarchical diagnosis of skin lesions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 2757–2765. [Google Scholar]
  30. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  31. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (ISBI), Hosted by the international skin imaging collaboration (ISIC). In Proceedings of the IEEE 15th International Symposium on Biomedical Imaging (ISBI), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
  32. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Halpern, A. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv 2019, arXiv:1902.03368. [Google Scholar]
  33. Nozdryn-Plotnicki, A.; Yap, J.; Yolland, W. Ensembling convolutional neural networks for skin cancer classification. In International Skin Imaging Collaboration (ISIC) Challenge on Skin Image Analysis for Melanoma Detection; MICCAI: Granada, Spain, 2018. [Google Scholar]
  34. Gessert, N.; Sentker, T.; Madesta, F.; Schmitz, R.; Kniep, H.; Baltruschat, I.; Schlaefer, A. Skin lesion diagnosis using ensembles, unscaled multi-crop evaluation and loss weighting. arXiv 2018, arXiv:1808.01694. [Google Scholar]
  35. Mahbod, A.; Schaefer, G.; Wang, C.; Dorffner, G.; Ecker, R.; Ellinger, I. Transfer learning using a multi-scale and multi-network ensemble for skin lesion classification. Comput. Methods Programs Biomed. 2020, 193, 105475. [Google Scholar] [CrossRef]
  36. Shen, S.; Xu, M.; Zhang, F.; Shao, P.; Liu, H.; Xu, L.; Xu, R.X. Low-cost and high-performance data augmentation for deep-learning-based skin lesion classification. arXiv 2021, arXiv:2101.02353. [Google Scholar]
  37. Xiao, F.; Wu, Q. Visual saliency based global–local feature representation for skin cancer classification. IET Image Process. 2020, 14, 2140–2148. [Google Scholar] [CrossRef]
  38. Yang, X.; Zeng, Z.; Yeo, S.Y.; Tan, C.; Tey, H.L.; Su, Y. A novel multi-task deep learning model for skin lesion segmentation and classification. arXiv 2017, arXiv:1703.01025. [Google Scholar]
  39. Zhang, J.; Xie, Y.; Xia, Y.; Shen, C. Attention Residual Learning for Skin Lesion Classification. IEEE Trans. Med Imaging 2019, 38, 2092–2103. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of skin lesions images; (a) actinic keratosis, (b) basal cell carcinoma, (c) benign keratosis-like, (d) dermatofibroma, (e) malignant melanoma, (f) melanocytic nevi, (g) vascular lesions.
Figure 1. Example of skin lesions images; (a) actinic keratosis, (b) basal cell carcinoma, (c) benign keratosis-like, (d) dermatofibroma, (e) malignant melanoma, (f) melanocytic nevi, (g) vascular lesions.
Applsci 13 05497 g001
Figure 2. Examples of border features; (a) original image, (b) border feature.
Figure 2. Examples of border features; (a) original image, (b) border feature.
Applsci 13 05497 g002
Figure 3. Example of a histogram for skin lesion image; (a) original image, (b) histogram of original image, (c) thresholding image, (d) histogram of thresholding image.
Figure 3. Example of a histogram for skin lesion image; (a) original image, (b) histogram of original image, (c) thresholding image, (d) histogram of thresholding image.
Applsci 13 05497 g003
Figure 4. Edge feature collected through inter & outer-border extraction and canny edge detector; (a) skin lesion image, i( · ), (b) thresholding image by [ q 1 , q 2 ] , I T · , (c) O s t u k image, I B O ( · ) , (d) O s t u k image, I B I ( · ) , (e) O s t u k edge image, I E O ( · ) , (f) O s t u k edge image, I E O ( · ) , (g) XOR image I E ( · ) of (e,f).
Figure 4. Edge feature collected through inter & outer-border extraction and canny edge detector; (a) skin lesion image, i( · ), (b) thresholding image by [ q 1 , q 2 ] , I T · , (c) O s t u k image, I B O ( · ) , (d) O s t u k image, I B I ( · ) , (e) O s t u k edge image, I E O ( · ) , (f) O s t u k edge image, I E O ( · ) , (g) XOR image I E ( · ) of (e,f).
Applsci 13 05497 g004
Figure 5. Classification of line-segment types using a L T D F ; (a) Some data from input Figure 4g, (b) responsive coefficient of L T D F , (c) responsive coefficient.
Figure 5. Classification of line-segment types using a L T D F ; (a) Some data from input Figure 4g, (b) responsive coefficient of L T D F , (c) responsive coefficient.
Applsci 13 05497 g005
Figure 6. Pattern-n feature of the response coefficient through the LTDF.
Figure 6. Pattern-n feature of the response coefficient through the LTDF.
Applsci 13 05497 g006
Figure 7. Image example of 3-color space; (a) RGB image, (b) YCrCb image, (c) HSV image.
Figure 7. Image example of 3-color space; (a) RGB image, (b) YCrCb image, (c) HSV image.
Applsci 13 05497 g007
Figure 8. Examples of texture extraction image T i , j ( · ) using LTEM mask M i , j .
Figure 8. Examples of texture extraction image T i , j ( · ) using LTEM mask M i , j .
Applsci 13 05497 g008
Figure 9. Example of extracted EFA[∙], CFA[∙], and TFA[∙] for skin lesion classification.
Figure 9. Example of extracted EFA[∙], CFA[∙], and TFA[∙] for skin lesion classification.
Applsci 13 05497 g009
Figure 10. A convolutional neural network model for learning a skin lesion. (a) Overview of a learning model. (b) A convolution group structure.
Figure 10. A convolutional neural network model for learning a skin lesion. (a) Overview of a learning model. (b) A convolution group structure.
Applsci 13 05497 g010
Figure 11. The Adagrad function (lr = 0.001) confusion matrix using the ISIC 2018 skin lesion datasets.
Figure 11. The Adagrad function (lr = 0.001) confusion matrix using the ISIC 2018 skin lesion datasets.
Applsci 13 05497 g011
Table 1. Line-segment type no. corresponding to the response of LTDF; EN: Eigen No., N-A: Non-activity, P: Point, H: Horizontality, V: Verticality, C: Curve, D: Diagonal, F: Face, P-n: Pattern-n.
Table 1. Line-segment type no. corresponding to the response of LTDF; EN: Eigen No., N-A: Non-activity, P: Point, H: Horizontality, V: Verticality, C: Curve, D: Diagonal, F: Face, P-n: Pattern-n.
No.FeatureNo.FeatureNo.FeatureNo.FeatureNo.FeatureNo.FeatureNo.FeatureNo.Feature
0N-A32P64P96H128P160P-1192H224H
1P33P-165P-497P-2129P-1161C193P-5225P-6
2P34P-466P-198P-2130P-4162P-7194P-2226P-12
3H35P-267P-299P-3131P-5163P-9195P-13227P-14
4P36P-168P-4100P-5132P-1164C196P-2228P-6
5P-137C69P-7101P-9133C165P-15197P-9229P-19
6H38P-570P-2102P-13134P-2166P-9198P-3230P-14
7H39P-671P-12103P-14135P-6167P-19199P-14231P-18
8P40V72D104C136P-4168P-2200C232C
9V41V73C105C137P-5169P-6201C233C
10D42C74C106C138P-7170P-8202P-9234C
11C43C75C107C139P-10171P-11203C235C
12P-444P-576P-2108P-10140P-7172P-9204P-8236P-11
13P-245P-677P-8109P-11141P-9173P-19205P-17237P-16
14C46C78P-9110C142P-8174P-17206C238C
15C47C79C111C143P-11175P-16207C239C
16P48P-480D112C144V176P-2208C240C
17P-449P-781P-7113P-8145P-5177P-9209P-10241P-11
18D50P-282C114P-9146C178P-8210C242C
19C51P-883P-9115C147C179P-17211C243C
20V52P-584C116C148V180P-6212C244C
21C53P-985P-8117P-17149P-6181P-19213P-11245P-16
22C54P-1086C118C150C182P-11214C246C
23C55P-1187C119C151C183P-16215C247C
24P-156P-288C120C152P-2184P-3216C248C
25P-257P-1289P-9121C153P-13185P-14217C249C
26C58P-990P-15122C154P-9186C218C250P-16
27C59C91C123P-16155C187C219P-18251C
28P-260P-1392P-9124C156P-12188P-14220C252C
29P-361P-1493C125C157P-14189P-18221C253C
30C62C94C126P-18158C190C222P-16254C
31C63C95P-16127C159C191C223C255F
Table 2. Visual line-segment type summary about Engen type in Table 1; TN: type no.
Table 2. Visual line-segment type summary about Engen type in Table 1; TN: type no.
TNTypeNo. of TypeTNTypeNo. of Type
1Non-Activity (N-A)114Pattern7 (P-7)8
2Point (P)815Pattern8 (P-8)4
3Verticality (V)616Pattern9 (P-9)8
4Horizontality (H)617Pattern10 (P-10)2
5Diagonal (D)418Pattern11 (P-11)17
6Curve (C)9719Pattern12 (P-12)8
7Face (F)120Pattern13 (P-13)4
8Pattern1 (P-1)621Pattern14 (P-14)4
9Pattern2 (P-2)822Pattern15 (P-15)4
10Pattern3 (P-3)1623Pattern16 (P-16)4
11Pattern4 (P-4)424Pattern17 (P-17)8
12Pattern5 (P-5)825Pattern18 (P-18)8
13Pattern6 (P-6)426Pattern19 (P-19)8
Table 3. Types of LTEM’s default masks.
Table 3. Types of LTEM’s default masks.
LabelTexture5-Value
1Level V 1 ( L 5 ) = [ 1 , 4 , 6 , 4 , 1 ]
2Edge V 2 ( E 5 ) = [ 1 , 2 , 0 , 2 , 1 ]
3Spot V 3 ( S 5 ) = [ 1 , 0 , 2 , 0 , 1 ]
4Wave V 4 ( W 5 ) = [ 1 , 2 , 0 , 2 , 1 ]
5Ripple V 5 ( R 5 ) = [ 1 , 4 , 6 , 4 , 1 ]
Table 4. Accuracy and loss of optimization ft. in proposed model on the ISIC 2018 datasets.
Table 4. Accuracy and loss of optimization ft. in proposed model on the ISIC 2018 datasets.
Optimization ft.Learning Rate (Accuracy/Loss)
0.010.0010.00010.00001
SGD0.85/0.85170.86/0.48710.84/0.44360.84/0.4436
RMSprop0.79/238.21690.80/14.80060.83/1.88110.85/0.8797
Adagrad0.88/0.58420.88/0.45430.85/0.42710.85/0.4097
Adadelta0.86/0.46550.83/0.44120.83/0.97280.79/0.6023
Adam0.82/1113.75720.82/20.21610.86/0.97280.86/0.8602
Adamax0.82/3.51220.84/1.25260.86/0.52420.86/0.9941
Nadam0.82/835.30740.83/22.93580.86/0.88730.84/0.4291
Table 5. Performance comparison result with a well-known CNN models for the ISIC 2018 dataset.
Table 5. Performance comparison result with a well-known CNN models for the ISIC 2018 dataset.
NetInput SizeAccuracyAvg. AUCREF
DenseNet-201512 × 51286.2%98.0%[10]
256 × 25682.0%97.5%
128 × 12875.2%95.0%
DenseNet-121512 × 51283.4%97.2%
256 × 25680.9%96.6%
128 × 12873.9%94.2%
SEResNeXt-101512 × 51285.7%98.1%
256 × 25683.1%97.8%
128 × 12874.5%96.0%
SEResNeXt-50512 × 51286.7%98.2%
256 × 25681.8%97.5%
128 × 12876.1%95.8%
ResNet-18512 × 51280.6%97.5%
256 × 25678.9%96.8%
128 × 12870.8%93.7%
ResNet-50512 × 51284.1%97.7%
256 × 25679.6%96.5%
128 × 12869.2%93.0%
ResNet-152512 × 51286.1%97.8%
256 × 25680.2%97.2%
128 × 12872.3%94.6%
Our Method50 × 25688.6%94.9%-
Table 6. Performance comparison result with a previous related references for the ISIC 2018 dataset.
Table 6. Performance comparison result with a previous related references for the ISIC 2018 dataset.
REFTeam/AuthorsAccuracySensitivity
for Melanoma
Avg.
AUC
Avg.
Specificity
[33]Nozdryn et al.88.5%76.0%98.3%83.3%
[34]Gassert et al.85.6%80.1%98.7%98.4%
[35]Amirreza et al.86.2%-98.1%-
[36]Shen et al.85.3%78.9%97.5%97.3%
[11]Zhuang et al.84.5%70.2%97.8%98.0%
Minjie89.5%77.8%98.2%98.1%
Amirreza et al.87.4%58.5%97.9%99.2%
MRCN87.2%86.0%95.3%97.9%
IPM_HPC86.6%83.0%97.6%97.6%
Mahbod83.6%71.9%97.5%98.2%
Yao, Peng, et al. (CMEL = 1.0)86.4%77.2%97.4%97.1%
Yao, Peng, et al. (CMEL = 1.8)85.2%86.0%97.3%96.3%
-Our Method88.6%71.7%94.9%97.1%
Table 7. Performance evaluation results of the proposed method and different DCNN models for the ISIC 2017 dataset; ACC (accuracy), SENS (sensitivity), SPEC (specificity).
Table 7. Performance evaluation results of the proposed method and different DCNN models for the ISIC 2017 dataset; ACC (accuracy), SENS (sensitivity), SPEC (specificity).
NetMelanoma (%)Seborrheic Keratosis (%)REF
ACCAUCSENSSPECACCAUCSENSSPEC
ResNet-5083.080.443.692.587.788.062.292.2[37]
DenseNet-12182.883.558.188.888.391.066.792.2
ResNet-50 & LBP83.878.452.188.690.590.972.291.3
DenseNet-121 & LBP85.883.261.591.791.793.472.295.1
Our Method89.387.189.089.386.391.686.086.3-
Table 8. Performance comparison result with a previous related references for the ISIC 2017 dataset; ACC (accuracy), SENS (sensitivity), SPEC (specificity).
Table 8. Performance comparison result with a previous related references for the ISIC 2017 dataset; ACC (accuracy), SENS (sensitivity), SPEC (specificity).
REFTeam/AuthorsMelanoma (%)Seborrheic Keratosis (%)
ACCAUCSENSSPECACCAUCSENSSPEC
[31]Codella, N.C. et al.83.083.043.692.591.794.270.099.5
[37]Xiao et al.-85.367.590.9-95.588.992.8
[38]Yang, X. et al.83.785.959.089.690.895.177.893.1
[39]Zhang, J. et al.85.883.261.591.797.793.472.295.1
-Our Method89.387.189.0%89.3%86.3%91.6%86.0%86.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, C.; Jang, M.; Han, Y.; Hong, Y.; Lee, W. Skin Lesion Classification Using Hybrid Convolutional Neural Network with Edge, Color, and Texture Information. Appl. Sci. 2023, 13, 5497. https://doi.org/10.3390/app13095497

AMA Style

Kim C, Jang M, Han Y, Hong Y, Lee W. Skin Lesion Classification Using Hybrid Convolutional Neural Network with Edge, Color, and Texture Information. Applied Sciences. 2023; 13(9):5497. https://doi.org/10.3390/app13095497

Chicago/Turabian Style

Kim, Changmin, Myeongsoo Jang, Younghwan Han, Yousik Hong, and Woobeom Lee. 2023. "Skin Lesion Classification Using Hybrid Convolutional Neural Network with Edge, Color, and Texture Information" Applied Sciences 13, no. 9: 5497. https://doi.org/10.3390/app13095497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop