Next Article in Journal
Mortality after Sustaining Skeletal Fractures in Relation to Age
Previous Article in Journal
The Role of Excessive Anticoagulation and Missing Hyperinflammation in ECMO-Associated Bleeding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Convolutional Neural Support Vector Machines for the Classification of Basal Cell Carcinoma Hyperspectral Signatures

by
Lloyd A. Courtenay
1,*,
Diego González-Aguilera
1,
Susana Lagüela
1,
Susana Del Pozo
1,
Camilo Ruiz
2,
Inés Barbero-García
1,
Concepción Román-Curto
3,4,
Javier Cañueto
3,4,5,
Carlos Santos-Durán
3,
María Esther Cardeñoso-Álvarez
3,
Mónica Roncero-Riesco
3,
David Hernández-López
6,
Diego Guerrero-Sevilla
6 and
Pablo Rodríguez-Gonzalvez
7
1
Department of Cartographic and Terrain Engineering, Higher Polytechnic School of Ávila, University of Salamanca, Hornos Caleros 50, 05003 Ávila, Spain
2
Department of Didactics of Mathematics and Experimental Sciences, Faculty of Education, Paseo de Canaleja 169, 37008 Salamanca, Spain
3
Department of Dermatology, University Hospital of Spain, Paseo de San Vicente 58-182, 37007 Salamanca, Spain
4
Instituto de Investigación Biomédica de Salamanca (IBSAL), Paseo de San Vicente 58-182, 37007 Salamanca, Spain
5
Institute of Molecular Biology and Cellular Cancer (IBMCC), Centre of Cancer Investigation (Lab 7), Campus Miguel de Unamuno s/n, 37007 Salamanca, Spain
6
Institute for Regional Development, University of Castilla la Mancha, Campus Universitario s/n, 02071 Albacete, Spain
7
Department of Mining Technology, Topography and Structures, University of León, Av. Astorga s/n, 24401 Ponferrada, Spain
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(9), 2315; https://doi.org/10.3390/jcm11092315
Submission received: 24 March 2022 / Revised: 18 April 2022 / Accepted: 19 April 2022 / Published: 21 April 2022
(This article belongs to the Section Dermatology)

Abstract

:
Non-melanoma skin cancer, and basal cell carcinoma in particular, is one of the most common types of cancer. Although this type of malignancy has lower metastatic rates than other types of skin cancer, its locally destructive nature and the advantages of its timely treatment make early detection vital. The combination of multispectral imaging and artificial intelligence has arisen as a powerful tool for the detection and classification of skin cancer in a non-invasive manner. The present study uses hyperspectral images to discern between healthy and basal cell carcinoma hyperspectral signatures. Upon the combined use of convolutional neural networks, with a final support vector machine activation layer, the present study reaches up to 90% accuracy, with an area under the receiver operating characteristic curve being calculated at 0.9 as well. While the results are promising, future research should build upon a dataset with a larger number of patients.

1. Introduction

Skin cancer, including melanoma and non-melanoma, is one of the most common types of cancer, especially among white-skinned and elderly populations, with incidence rates still on the rise [1,2,3,4]. Although cases of non-melanoma skin cancer (NMSC) present a lower mortality rate than melanoma, its incidence rate is up to 20 times higher. The most common types of NMSC are basal cell carcinoma (BCC) and cutaneous squamous cell carcinoma (SCC or cSCC). BCC represents the majority of cases, while the steady increase in BCC patients suggests that this type of cancer may prevail over other types of skin cancer combined [3].
Skin malignancies are firstly evaluated by visual screening. For suspicious lesions, this evaluation is usually followed by biopsy and histopathological analyses of skin tissues. The high importance of early diagnosis has encouraged the development of methodologies for the automatic and more efficient detection of these skin lesions. This process is a challenging task, however, due to the large variability of skin lesions. Nevertheless, computational learning (CL) strategies, such as machine and deep learning, have emerged as powerful tools for these sorts of classification problems [5].
Several studies focus on the capacities of CL for skin cancer detection and classification [5,6]. In spite of this, not all pathologies are covered according to their importance and prevalence (Table 1). While recent years have seen an increase in the number of publications focusing on NMSCs, cancer types such as BCC, which are by far the most common [7], have received relatively little attention in the field of artificial intelligence (AI). Early detection of BCC is important considering the locally destructive nature of the disease and its increased depth of invasion, as well as difficulties in treatment for more severe cases [8,9]. Moreover, BCCs can present a possible increase in metastasis rates, with an increase of 1–2% if treatment is delayed for tumors >3 cm in diameter and an increase of up to 50% for tumors <10 cm [10]. From this perspective, the development of tools for tumor detection and analysis may speed up clinical practice by presenting a new means of planning treatments, and interventions, as well as studying lesions.
Computer vision has proven to have great potential for the non-invasive detection and classification of skin cancer lesions. While some methodologies have proven useful using conventional cameras in the visible range of the electromagnetic spectrum [5,11,12], hyperspectral imagery is also widely used [13]. Hyperspectral cameras have a high number of bands or channels, usually with wavelengths beyond the visible areas of the spectrum. Such sensors provide a large quantity of information that can be used for tumor identification and classification [14,15]. Nevertheless, hyperspectral images present the distinct disadvantage of being very large and hard to process. From this perspective, a very large dataset would be required for the efficient processing of these images using most CL techniques.
A recent study analyzed the spectral information provided by near-infrared hyperspectral images [16]. Using a pixel-based approach, said study employed robust statistical tests for feature selection. The authors were able to identify an optimal electromagnetic window that can be used to distinguish between different types of NMSC, as well as healthy skin. The present study develops this, employing this pixel-based approach to find an optimal neural network architecture for the classification of hyperspectral signatures.
In the field of applied CL in computer vision, convolutional neural networks (CNNs) can be considered a state-of-the-art algorithm for most image processing applications [17]. Part of this success is due to the wide-spread research into optimal CNN architectures, including ImageNet [18], VGG [19], ResNet [20], and Inception [21]. In light of this, CNNs have also been found useful for the processing of vectorial data [22], such as raw audio-waveforms [23], seismic vibrations [24], and electrocardiogram data [25]. CNNs can thus be considered a versatile set of algorithms, being adept at the extraction of spatially or temporally invariant features while reaching high performance in competition with, or even surpassing, human specialists [17].
Other algorithms such as support vector machines (SVMs) have also proved highly efficient for the processing of complex data types. Useful for both regression and discriminant analyses, SVMs use a kernel trick to overcome traditional limitations imposed by data linearity and normality [26]. From this perspective, SVMs have proven highly useful for the processing of different types of information, including some applications in computer vision [13,27], natural language processing [28,29], material quality inspection [30], and geometric morphometrics [31].
The concept of a neural support vector machine (NSVM) was originally proposed as a means of adding “depth” to SVMs [32], while using non-linear neural network architectures as a means of training a specified kernel function directly on the dataset. The added use of neural networks thus adds a highly versatile and flexible “kernel” for the SVM. Applications inspired by this approach have found NSVMs to be highly useful for both regression and classification tasks [33], extracting high-level features from low-level domains [34]. NSVMs have since proven successful in the processing of geometric data derived from 3D models [31], as well as other promising approaches in the classification of hyperspectral images [35].
The present study builds on each of the predefined approaches, using a convolutional NSVM (CNSVM) for the classification of BCC hyperspectral signatures. The present architecture employs 1D inception modules for convolution over hyperspectral data, performing feature extraction and thus acting as a kernel for the SVM activation layer. Both the network and the final SVM activation layer are fine-tuned using Bayesian approaches, reaching up to 90% overall accuracy in differentiating between healthy and cancerous tissue. Upon defining this architecture, future research including larger sample sizes may be a strong starting point for the classification and segmentation of entire hyperspectral images, facilitating diagnosis, patient screening, and the delimitation of cancerous regions.

2. Materials and Methods

2.1. Sample

The data set used in the present study was derived from previous works by Courtenay et al. [16], available online from [36]. This dataset consists of a total of 1505 hyperspectral signatures of three different samples, including BCC, cSCC, and healthy (H) skin samples (Figure 1).
Hyperspectral signatures for each patient were obtained using a Headwall Nano-Hyperspec visible–near-infrared (VNIR) hyperspectral imaging sensor. This particular sensor is a pushbroom linear camera, producing a vectorial array of pixels (1 × 640 px), registering wavelengths between 398.08 and 995.20 nm with a 2.2 nm spectral bandwidth separation between channels. The sensor was fit onto an ad hoc platform controlled by an electronic module device, designed for the synchronization of the platform’s movement, illumination, and the sensor’s shutter speed. Calibration of the sensor was performed using a marker board and frame presenting a known reflectance pattern (Spectralon). Pixel values were then calculated and radiometrically corrected through this calibration process so as to produce reflectance values (in%) for each channel of each pixel. Final images thus consisted of a 431 × 851 × 260 (rows, columns, and channels) tensor (Figure 1), having ≈95.4 million numeric values and occupying ≈0.3 GB of memory. For more details about the image acquisition process, consult the work of Courtenay et al. [16].
For the purpose of processing and characterizing hyperspectral signatures, Courtenay et al. [16] defined regions of interest (ROIs) for each of the images, randomly sampling pixels from ROIs directly over the tumor, while H samples were extracted at ROIs furthest away from the tumor so as to avoid possible contamination. A total of 41 BCC patients were originally studied. While the dataset includes data from cSCC patients, these signatures originate from a much smaller number of patients and were thus excluded in the present study. In light of this, the final sample size included here consists of 504 BCC signatures and 488 H signatures, amounting to a total sample size of 992. Considering the presence of only two labels (BCC and H), the training of all deep learning algorithms was thus conceptualized as a supervised binary classification problem, discerning whether the hyperspectral signatures represented cancer (BCC = 1) or not (H = 0).
All patients agreed to participate in the study; however, due to patient anonymity and data protection, no further details have been disclosed.

2.2. Base Convolutional Neural Network Architecture

The present architecture takes as input hyperspectral signatures represented as an ℝn first-order tensor, where n represents the number of hyperspectral channels included for the characterization of each sample. The previous publication of this dataset employed robust statistical approaches to define an optimal window between 573.45 and 779.88 nm (Figure 1). This region was defined as the portion of the electromagnetic spectrum where statistical differences between samples are most likely to be present [16]. Under this premise, all hyperspectral signatures were cropped to only include wavelengths within this window, producing a final input of ℝ94.
The convolutional portion of the network is inspired by the Inception architecture, typically used in computer vision applications [21]. For this purpose, the inception module was adapted for 1D convolutions using a block of parallel convolutional layers of varying receptive field sizes (1 × 1, 1 × 3 or 1 × 5), as well as different numbers of filters per layer (Figure 2). Padding was used for all filters, while filter strides were set to 1 for all layers. Within the inception module, and parallel to these convolutional filters, an additional 3 × 1 max-pooling layer was also included. The output of the max-pooling layer was also passed into a convolutional filter (Figure 2).
After each convolutional layer, batch normalization was used prior to activation, while two different non-linear activation functions were tried and tested (Equations (1) and (2), Figure 3). These were the rectified linear unit (ReLU) (Equation (1)), and the self-gated rectified activation function (Swish) (Equation (2)) [37];
f ( x ) = m a x ( 0 , x )
f ( x ) = x · 1 1 + e x
Additional hyperparameter configurations considered the use of kernel initializers and regularizers in each of the convolutional layers. The best results were obtained using the LeCun normal initializer [38,39], as well as an ℓ2 regularizer with a coefficient of 0.0001 [40]. The results of each module were concatenated before being passed on to the next portion of the algorithm.
Experiments were performed by stacking a different number of inception modules on top of each other, as well as employing the use of only a single inception module prior to the fully connected neural networks that followed. While fully convolutional networks were also experimented with, considering their success in other applications [23], the present study found these architectures to considerably over- or under-fit on our training data.
Following the convolutional layers of our model, the concatenated output was flattened into a large vector and subsequently passed into a dense fully connected neural network. Different experiments considered the size and density of these fully connected layers, while a final aggressive reduction tactic was employed, ensuring each subsequent layer to be at least half the size of its predecessor. In between each of these layers, dropout algorithms were inserted for a more efficient training. Similarly to the convolutional portion of our algorithm, both ReLU and Swish activation functions were considered [37]. In the original configuration of the neural network, the final layer consisted of a single sigmoid activated neuron.
Training used a train:validation split of 70:30%, while training data were shuffled for each epoch. In these initial experiments, no data augmentation techniques were performed.
CNN models were trained using either a vanilla stochastic gradient descent (SGD) algorithm, or the Adam optimization algorithm [41]. Cyclic learning rates (CLRs) were employed [42], using the Triangular2 function (base α = 0.0001, max α = 0.01, step size = 8). Models were run for 100 to 500 epochs until convergence, using batch sizes of 16, 32, 64, and 128. Finally, algorithms were implemented in the Python v.3.7.2 programming language, using the Tensorflow 2.1.0 library [43], and the TensorBoard toolkit for callback evaluation during training.

2.3. Convolutional Neural Support Vector Machines

Once the base architecture of the CNN had been trained, the final sigmoid activated layer was removed, replaced, and retrained with an SVM activation layer. For this purpose, each hyperspectral signature was passed through the CNN for feature extraction and then used to train the SVM [31]. For the tuning of this layer, k-fold cross-validation (k = 10) was used, while additional experiments alternated between a linear, polynomial, or radial SVM kernel functions.
The SVM portion of our network was implemented using the Scikit Learn v.0.22.1 library [44].

2.4. Hyperparameter Optimization

Hyperparameter optimization for multiple components of each model was performed using Bayesian optimization algorithms (BOAs) [45,46]. For the CNN component of our algorithms, BOAs were used to define targeted values such as the optimal dropout threshold, optimal parameters of ℓ2 regularization, optimal neural layer densities, and optimal learning rates for CLR tuning. Similarly, for the SVM portion of our model, BOAs were used to define the optimal coefficients used in the kernel function. For all experiments, BOAs were run for 100 iterations, using an expected improvement selection function via the Tree of Parzen Estimator algorithm [46].

2.5. Model Evaluation

Six different performance indices were recorded for the evaluation of the present architecture, calculated on the test set for each of the experiments. The first of these included overall accuracy, sensitivity (i.e., true positive rate), and specificity (i.e., true negative rate) [47]. While other studies frequently use additional calculations such as precision and recall, these metrics are more suitable for studies presenting class imbalance [48]. Considering how this is not the case here, these metrics were thus excluded. Alongside specificity and sensitivity calculations, receiver operating characteristic (ROC) curves were also calculated, with their corresponding area under curve (AUC) values [49]. In addition to this, the kappa (κ) statistic was also used to assess the probability of agreement between the predicted output and the original label [47,50,51]. Finally, a calculation of the model’s final loss was performed using the mean squared error (MSE) metric.
For the evaluation of these metrics, a model was considered powerful if specificity and sensitivity values appeared balanced (sensitivity ≈ specificity), thus implicating high AUC values (<0.8). For kappa statistics, κ < 0.8 was considered as a threshold for almost perfect agreement between the real and predicted labels [47,50,51]. Finally, overall accuracy was only considered a reliable metric if all of these criteria had been met.

2.6. Computational Facilities

For experimental purposes, three different computer systems were used for training, taking note of training time, CPU efficiency, and RAM usage. This was performed to assess not only the computational cost of the algorithms, but also their replicability on any standard computer system.
The first of these systems was a Dell Precision T1700 desktop computer, equipped with an Intel Xeon E3-1240 v3 CPU processor (4 cores, 3.40 GHz operating frequency) and 8 GB of RAM. The second system was a portable ASUS X550VX portable laptop, equipped with an Intel Core i5-6300HQ CPU processor (4 cores, 2.30 GHz) and 8 GB of RAM. Finally, the services provided by the Supercomputation Center of Castilla y Leon (SCAYLE) were also experimented with, employing the use of the Broadwell architecture for CL. This SCAYLE server has 2 Intel Xeon E5-2695 v4 CPU processors (18 cores each, 2.10 GHz), and 384 GB of RAM. In addition to this, this SCAYLE server provides access to a total of 8 NVidia Tesla v100 GPUs. For this purpose, experiments were performed using different numbers of CPU cores, as well as the use of GPUs, for training and data processing.

3. Results

Of all the configurations tried and tested, the model architecture that was found to produce the best results consisted of two 1D inception modules, feeding into a four-layer neural network, with densities of 200, 150, 100, and 50 neurons respectively (Table 2). This resulted in a final model of ≈5 million trainable parameters. Across all cases, models were found to converge after 400 epochs, while batch sizes of 128 provided algorithms with enough information to successfully train from. CNN architectures by themselves were not reported to produce promising results for the classification of hyperspectral signatures, reaching an accuracy of only approximately 56.7%, AUC of 0.61, and κ values as low as 0.3. Nevertheless, with the inclusion of the final SVM activation layer, CNSVM algorithms presented between a 12.1% and a 49.3% improvement in performance (Figure 4), with the radial SVM algorithm reaching the highest recorded results (accuracy = 91.0%, κ = 0.81, AUC = 0.91). Table 2 presents the final CNSVM architecture used in this study.
When fine-tuning the CNSVM considering the different types of activation functions and optimization algorithms, both Swish and ReLU were observed to produce optimal results (Table 3, Figure 5). Nevertheless, Adam was observed to fit the Swish models better than ReLU, while SGD produced the best results on all accounts (Table 3, Figure 5). Finally, MSE results reveal all models to produce confident predictions, with the Swish and SGD algorithm reaching the best results, as seen by a mean of 97.1% confidence with each prediction made on the test set.
Finally, when considering the computational cost of these algorithms, as would be expected, training time was considerably conditioned by the inclusion of a GPU (Table 4), while the higher number of CPU cores also reduced training times. The longest recorded training time was registered at 40 min when using the personal laptop (Table 4), with each epoch taking 5.9 s. The inclusion of a single GPU was found to train neural networks 24 times faster. The time taken to tune the SVM layer via BOAs was not seen to be affected by the number of CPUs, with all computer systems taking ≈1.7 min to tune the final activation layer and 0.01 min to fit. As for computational resources, all computers required ≈1.5 GB of RAM for the training of models, while CPU efficiency was recorded at 80%. In sum, the CNSVM for hyperspectral signature classification can easily be trained without the need for high computational facilities.

4. Discussion

The present study proposes a convolutional neural support vector machine architecture for the classification of healthy skin and basal cell carcinoma hyperspectral signatures. As can be seen, the network works well on test sets, reaching up to 90% classification in terms of overall accuracy, producing highly confident predictions, and is relatively easy to train. Multiple experiments throughout this paper have found both the Swish and ReLU activation functions to perform well, with ReLU performing marginally better in some applications. Finally, the results presented here can be considered an additional example of how support vector machine activation layers can prove highly effective when combined with deep neural network architectures.
Preliminary use of the CNSVM model to classify every pixel in an image proved quite successful (Figure 6A). Nevertheless, some issues were found in images presenting poorer lighting conditions (Figure 6B). In these latter cases, it can be seen how areas of shadow that are typically found in the creases and wrinkles of many patients’ faces evidently reflect less light. This feature unfortunately is what predominantly separates BCC from healthy skin [16]. Similarly, preliminary observations noted how BCC presents a large inter-patient variability, implying that a larger sample of cancer patients should definitely be considered in the future.
While each of these limitations are noteworthy, these can be overcome through the collection of larger datasets that will better represent the variability of the lesions, as well as the incorporation of improvements to the lighting setup of the sensor’s platform (as noted by [16]). Moreover, CNSVMs still proved to be highly successful when learning from this data, showing how future applications employing transfer learning [52,53], or developed generative data augmentation techniques [31,54,55,56], are still likely to perform well. Each of these steps must be considered fundamental before applications in real clinical settings.
The goal of this study was to develop previous observations regarding the differences observed via robust statistical techniques [16]. Here we have shown how advanced computational techniques are able to effectively learn these differences, precisely in the spectral range between 573.45 and 779.88 nm. Linked with other research in the field of hyperspectral imagery and dermatological analyses, other regions of the electromagnetic spectrum may be considered useful, branching out further into the near-infrared (1000 to 1700 nm) or short-wave infrared (1000 to 2500 nm) portion of the spectrum. This is especially relevant for the study of some types of skin cancer, such as melanoma [13] and cSCC [57].
Needless to say, CNSVMs can be considered a valuable tool for the processing of this type of data, presenting an important framework to build upon for more developed and advanced applications in real clinical settings.

Author Contributions

L.A.C., conceptualization, methodology, software, validation, formal analysis, investigation, data curation, writing—original draft, review and editing, visualization. D.G.-A., investigation, resources, writing—review and editing, supervision, project administration, funding acquisition. S.L., investigation, supervision, project administration, funding acquisition. S.D.P., investigation. C.R., investigation. I.B.-G., writing—original draft, review and editing. C.R.-C., resources, supervision, project administration, funding acquisition. J.C., investigation. C.S.-D., investigation. M.E.C.-Á., investigation. M.R.-R., investigation. D.H.-L., supervision. D.G.-S., formal analysis, investigation. P.R.-G., supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This project was funded by the Junta de Castilla y Leon, under the project title HYPER-SKINCARE (Ref. GRS 1837/A/18). Lloyd Austin Courtenay is funded by the Spanish Ministry of Science, Innovation and Universities with an FPI Predoctoral Grant (Ref. PRE2019-089411) associated with project RTI2018-099850-B-I00 and the University of Salamanca. Susana Lagüela and Susana Del Pozo are both funded by the University of Salamanca and Iberdrola Spain through the initiative Cátedra Iberdrola VIII Centenario of the University of Salamanca. Javier Cañueto is partially supported by PI18/00587 (Instituto de Salud Carlos III cofinanciado con fondos FEDER) and GRS 2139/A/20 (Gerencia Regional de Salud de Castilla y León).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Investigation with Medicine in the Health Sector of Salamanca (CEIm code: PI 2019 01 177, 6 February 2019).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

All data used in the current study are located at the corresponding author’s GitHub repository: https://github.com/LACourtenay/HyperSkinCare_Statistics (accessed on 2 September 2021).

Acknowledgments

This research has made use of the high-performance computing resources of the Castilla y León Supercomputing Center (SCAYLE, www.scayle.es (accessed on 18 April 2022)), financed by the European Regional Development Fund (ERDF).

Conflicts of Interest

The corresponding author has no conflict of interest to declare.

Code Availability

All code used in the current study is located at the corresponding author’s GitHub repository: https://github.com/LACourtenay/HyperSkinCare_DL (accessed on 2 September 2021).

References

  1. Brunssen, A.; Waldmann, A.; Eisemann, N.; Katalinic, A. Impact of skin cancer screening and secondary prevention campaigns on skin cancer incidence and mortality: A systematic review. J. Am. Acad. Dermatol. 2017, 76, 129–139.e10. [Google Scholar] [CrossRef] [PubMed]
  2. Lomas, A.; Leonardi-Bee, J.; Bath-Hextall, F. A systematic review of worldwide incidence of nonmelanoma skin cancer. Br. J. Dermatol. 2012, 166, 1069–1080. [Google Scholar] [CrossRef] [PubMed]
  3. Madan, V.; Lear, J.T.; Szeimies, R.M. Non-melanoma skin cancer. Lancet 2010, 375, 673–685. [Google Scholar] [CrossRef] [Green Version]
  4. Rogers, H.W.; Weinstock, M.A.; Harris, A.R.; Hinckley, M.R.; Feldman, S.; Fleischer, A.B.; Coldiron, B.M. Incidence Estimate of Nonmelanoma Skin Cancer in the United States, 2006. Arch. Dermatol. 2010, 146, 283–287. [Google Scholar] [CrossRef] [PubMed]
  5. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  6. Fujisawa, Y.; Otomo, Y.; Ogata, Y.; Nakamura, Y.; Fujita, R.; Ishitsuka, Y.; Watanabe, R.; Okiyama, N.; Ohara, K.; Fujimoto, M. Deep learning based, computer aided classifier developed with a small dataset of clinical images surpasses board certified dermatologists in skin tumour diagnosis. Br. J. Dermatol. 2019, 180, 373–381. [Google Scholar] [CrossRef]
  7. Diepgen, T.L.; Mahler, V. The epidemiology of skin cancer. Br. J. Dermatol. 2002, 146, 1–6. [Google Scholar] [CrossRef]
  8. Tillman, E.; Parekh, P.K.; Grimwood, R.E. Locally destructive metastatyic basal cell carcinoma. Cutis 2019, 103, E23–E25. [Google Scholar]
  9. Millan-Cayetano, J.-F.; Blazquez-Sanchez, N.; Fernandez-Canedo, I.; Repiso-Jiménez, J.B.; Funez-Liebana, R.; Bautista, M.D.; De Troya-Martin, M. Metastatic Basal Cell Carcinoma: Case Report and Review of the Literature. Indian J. Dermatol. 2020, 65, 61–64. [Google Scholar] [CrossRef]
  10. Hoorens, I.; Vossaert, K.; Ongenae, K.; Brochez, L. Is early detection of basal cell carcinoma worthwhile? Systematic review based on the WHO criteria for screening. Br. J. Dermatol. 2016, 174, 1258–1265. [Google Scholar] [CrossRef]
  11. Dai, X.; Spasic, I.; Meyer, B.; Chapman, S.; Andres, F. Machine Learning on Mobile: An On-Device Inference App for Skin Cancer Detection. In Proceedings of the 2019 Fourth International Conference on Fog and Mobile Edge Computing (FMEC), Rome, Italy, 10–13 June 2019; pp. 301–305. [Google Scholar] [CrossRef]
  12. Zhang, N.; Cai, Y.-X.; Wang, Y.-Y.; Tian, Y.-T.; Wang, X.-L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 101756. [Google Scholar] [CrossRef] [PubMed]
  13. Leon, R.; Martinez-Vega, B.; Fabelo, H.; Ortega, S.; Melian, V.; Castaño, I.; Carretero, G.; Almeida, P.; Garcia, A.; Quevedo, E.; et al. Non-Invasive Skin Cancer Diagnosis Using Hyperspectral Imaging for In-Situ Clinical Support. J. Clin. Med. 2020, 9, 1662. [Google Scholar] [CrossRef] [PubMed]
  14. Johansen, T.H.; Møllersen, K.; Ortega, S.; Fabelo, H.; Garcia, A.; Callico, G.M.; Godtliebsen, F. Recent advances in hyperspectral imaging for melanoma detection. WIREs Comput. Stat. 2020, 12, 1465. [Google Scholar] [CrossRef]
  15. Kuzmina, I.; Diebele, I.; Jakovels, D.; Spigulis, J.; Valeine, L.; Kapostinsh, J.; Berzina, A. Towards noncontact skin melanoma selection by multispectral imaging analysis. J. Biomed. Opt. 2011, 16, 060502. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Courtenay, L.A.; González-Aguilera, D.; Lagüela, S.; del Pozo, S.; Ruiz-Mendez, C.; Barbero-García, I.; Román-Curto, C.; Cañueto, J.; Santos-Durán, C.; Cardeñoso-Álvarez, M.E.; et al. Hyperspectral imaging and robust statistics in non-melanoma skin cancer analysis. Biomed. Opt. Express 2021, 12, 5107. [Google Scholar] [CrossRef]
  17. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, UK, 2016. [Google Scholar]
  18. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2012, 60, 84–90. [Google Scholar] [CrossRef]
  19. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; Available online: https://arxiv.org/pdf/1409.1556.pdf (accessed on 2 September 2021).
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar] [CrossRef] [Green Version]
  22. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  23. Dai, W.; Dai, C.; Qu, S.; Li, J.; Das, S. Very deep convolutional neural networks for raw waveforms. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, LA, USA, 5–9 March 2017; pp. 421–425. [Google Scholar] [CrossRef] [Green Version]
  24. Woollam, J.; Rietbrock, A.; Bueno, A.; De Angelis, S. Convolutional Neural Network for Seismic Phase Classification, Performance Demonstration over a Local Seismic Network. Seism. Res. Lett. 2019, 90, 491–502. [Google Scholar] [CrossRef]
  25. Acharya, U.R.; Fujita, H.; Lih, O.S.; Hagiwara, Y.; Tan, J.H.; Adam, M. Automated detection of arrhythmias using different intervals of tachycardia ECG segments with convolutional neural network. Inf. Sci. 2017, 405, 81–90. [Google Scholar] [CrossRef]
  26. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  27. Csurka, G.; Dance, C.; Fan, L.; Willamowski, J.; Bray, C. Visual categorization with bags of keypoints. Workshop Stat. Learn. Comput. Vis. 2004. Available online: https://www.cs.cmu.edu/~efros/courses/LBMV07/Papers/csurka-eccv-04.pdf (accessed on 20 April 2022).
  28. Barber, E.L.; Garg, R.; Persenaire, C.; Simon, M. Natural language processing with machine learning to predict outcomes after ovarian cancer surgery. Gynecol. Oncol. 2020, 160, 182–186. [Google Scholar] [CrossRef] [PubMed]
  29. Hao, P.-Y.; Kung, C.-F.; Chang, C.-Y.; Ou, J.-B. Predicting stock price trends based on financial news articles and using a novel twin support vector machine with fuzzy hyperplane. Appl. Soft Comput. 2021, 98, 106806. [Google Scholar] [CrossRef]
  30. Rodríguez-Martín, M.; Fueyo, J.; Gonzalez-Aguilera, D.; Madruga, F.; García-Martín, R.; Muñóz, A.; Pisonero, J. Predictive Models for the Characterization of Internal Defects in Additive Materials from Active Thermography Sequences Supported by Machine Learning Methods. Sensors 2020, 20, 3982. [Google Scholar] [CrossRef]
  31. Courtenay, L.A.; Herranz-Rodrigo, D.; González-Aguilera, D.; Yravedra, J. Developments in data science solutions for carnivore tooth pit classification. Sci. Rep. 2021, 11, 10209. [Google Scholar] [CrossRef]
  32. Wiering, M.A.; Ree, M.H.; Embrechts, M.J.; Stollenga, M.F.; Meijster, A.; Nolte, A.; Schomaker, L.R.B. The Neural Support Vector Machine. In Proceedings of the 25th Benelux Artificial Intelligence Conference, Delft, The Netherlands, 7–8 November 2013; pp. 254–257. [Google Scholar]
  33. Rahimi, A.; Recht, B. Random features for large-scale kernel machines. Adv. Neural Inf. Process. Syst. 2007, 20, 1–8. [Google Scholar]
  34. Tancik, M.; Srinivasan, P.P.; Mildenhall, B.; Fridovich-Keil, S.; Raghavan, N.; Singhal, U.; Ramamoorthi, R.; Barron, J.; Ng, R. Fourier features let networks learn high frequency functions in low dimensional domains. arXiv 2020, arXiv:2006.10739. Available online: https://arxiv.org/pdf/2006.10739v1.pdf (accessed on 2 September 2021).
  35. Okwuashi, O.; Ndehedehe, C.E. Deep support vector machine for hyperspectral image classification. Pat. Recog. 2020, 103, 107298. [Google Scholar] [CrossRef]
  36. Courtenay, L.A. Code and Data for the HYPER-SKINCARE project and paper titled ‘Hyperspectral Imaging and Robust Statistics in Non-Melanoma Skin Cancer Analysis’. GitHub 2021. Available online: https://github.com/LACourtenay/HyperSkinCare_Statistics (accessed on 2 September 2021).
  37. Ramachandran, P.; Zoph, B.; Le, Q.V. Swish: A self-gated activation function. Google Brain 2017, 1–12. Available online: https://arxiv.org/pdf/1710.05941v1.pdf (accessed on 2 September 2021).
  38. LeCun, Y.; Bottou, L.; Orr, G.B.; Müller, K.R. Efficient Back Prop. 1998. Available online: http://yann.lecun.com/exdb/publis/pdf/lecun-98b.pdf (accessed on 2 September 2021).
  39. Klambauer, G.; Unterthiner, T.; Mayr, A. Self-normalizing neural networks. Conf. Neur. Info. Process. Syst. 2017, 31, 972–981. [Google Scholar] [CrossRef]
  40. Krogh, A.; Hertz, J.A. A simple weight decay can improve generalization. Adv. Neur. Info. Process. Syst. 1991, 4, 950–957. [Google Scholar]
  41. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference Learn Represent, San Diego, CA, USA, 5–8 May 2015; Available online: https://arxiv.org/pdf/1412.6980.pdf (accessed on 2 September 2021).
  42. Smith, L.N. Cyclical learning rates for training neural networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 464–472. [Google Scholar] [CrossRef] [Green Version]
  43. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Tensorflow: Large-Scale Machine Learning on Heterogeneous Systems. arXiv 2015, arXiv:1603.04467. Available online: https://www.tensorflow.org/ (accessed on 1 September 2021). [CrossRef]
  44. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  45. Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. In Proceedings of the International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25, pp. 2951–2959. Available online: https://arxiv.org/pdf/1206.2944.pdf (accessed on 2 September 2021).
  46. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for hyper-parameter optimization. In Proceedings of the International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 12–15 December 2011; Volume 24, pp. 2456–2554. [Google Scholar] [CrossRef]
  47. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: Dordrecht, The Netherlands, 2013. [Google Scholar]
  48. He, H.; Ma, Y. Imbalanced Learning: Foundations, Algorithms and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  49. Sing, T.; Sander, O.; Beerenwinkel, N.; Lengauer, T. ROCR: Visualizing classifier performance in R. Bioinformatics 2005, 21, 3940–3941. [Google Scholar] [CrossRef]
  50. Landis, J.R.; Koch, G.G. The Measurement of Observer Agreement for Categorical Data. Biometrics 1977, 33, 159–174. [Google Scholar] [CrossRef] [Green Version]
  51. Foody, G.M. Harshness in image classification accuracy assessment. Int. J. Remote Sens. 2008, 29, 3137–3158. [Google Scholar] [CrossRef] [Green Version]
  52. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  53. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Neural Information Processing Systems, Montréal, QC, Canada, 8–13 December 2014; pp. 3320–3328. Available online: https://arxiv.org/pdf/1411.1792v1.pdf (accessed on 2 September 2021).
  54. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. In Proceedings of the International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 8–13 December 2014; pp. 2672–2680. Available online: https://arxiv.org/pdf/1406.2661v1.pdf (accessed on 2 September 2021).
  55. Lucic, M.; Kurach, K.; Michalski, M.; Bousquet, O.; Gelly, S. Are GANs created equal? A large scale study. In Proceedings of the 32nd Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Available online: https://arxiv.org/pdf/1711.10337v4.pdf (accessed on 2 September 2021).
  56. Courtenay, L.A.; González-Aguilera, D. Geometric morphometric data augmentation using generative computational learning algorithms. Appl. Sci. 2020, 10, 9133. [Google Scholar] [CrossRef]
  57. Brouwer de Koning, S.G.; Weijtmans, P.; Karakullukcu, M.B.; Shan, C.; Baltussen, E.J.M.; Smit, L.A.; van Veen, R.L.P.; Hendriks, B.H.W.; Sterenborg, H.J.C.M.; Ruers, T.J.M. Toward assessment of resection margins using hyperspectral diffuse reflection imaging (400–1700 nm) during tongue cancer surgery. Lasers Surg. Med. 2020, 52, 496–502. [Google Scholar] [CrossRef]
Figure 1. Examples of the hyperspectral signatures and images of healthy skin and basal cell carcinoma tumors.
Figure 1. Examples of the hyperspectral signatures and images of healthy skin and basal cell carcinoma tumors.
Jcm 11 02315 g001
Figure 2. Figurative schematic representing the architecture of the 1D Inception modules used in the present study. Convolutional filters are described by [N° filters, receptive field (rows × columns)]. Batch Norm. indicates batch normalization, while activation layers depend on the configuration of the algorithm at the time of training.
Figure 2. Figurative schematic representing the architecture of the 1D Inception modules used in the present study. Convolutional filters are described by [N° filters, receptive field (rows × columns)]. Batch Norm. indicates batch normalization, while activation layers depend on the configuration of the algorithm at the time of training.
Jcm 11 02315 g002
Figure 3. Graphical representation of the rectified linear unit (ReLU) and the self-gated rectified (Swish) activation functions (f(x)), alongside their first (f′(x)) and second (f″(x)) derivatives.
Figure 3. Graphical representation of the rectified linear unit (ReLU) and the self-gated rectified (Swish) activation functions (f(x)), alongside their first (f′(x)) and second (f″(x)) derivatives.
Jcm 11 02315 g003
Figure 4. Receiver operating characteristic curves alongside their calculated area under curve (AUC) statistics for the different support vector machine activations used. CNN = base convolutional neural network without support vector machine activation. CNSVM = convolutional support vector machine.
Figure 4. Receiver operating characteristic curves alongside their calculated area under curve (AUC) statistics for the different support vector machine activations used. CNN = base convolutional neural network without support vector machine activation. CNSVM = convolutional support vector machine.
Jcm 11 02315 g004
Figure 5. Radar plots comparing performance metrics of each of the configurations tried and tested. AUC = area under curve. ReLU = rectified linear unit. SGD = stochastic gradient descent. The red line at 0.8 marks a suitable threshold defining an optimal computational learning model.
Figure 5. Radar plots comparing performance metrics of each of the configurations tried and tested. AUC = area under curve. ReLU = rectified linear unit. SGD = stochastic gradient descent. The red line at 0.8 marks a suitable threshold defining an optimal computational learning model.
Jcm 11 02315 g005
Figure 6. Preliminary examples of (A) good and (B) poor image segmentation using CNSVMs for the classification of each pixel. (A) Examples of BCC tumors found on the forehead of a male patient and shoulder of a female patient. (B) Examples of BCC tumors found in the crease between the cheek and nostril of two female patients. Due to patient anonymity, images have been cropped to avoid revealing any distinguishing features.
Figure 6. Preliminary examples of (A) good and (B) poor image segmentation using CNSVMs for the classification of each pixel. (A) Examples of BCC tumors found on the forehead of a male patient and shoulder of a female patient. (B) Examples of BCC tumors found in the crease between the cheek and nostril of two female patients. Due to patient anonymity, images have been cropped to avoid revealing any distinguishing features.
Jcm 11 02315 g006
Table 1. A bibliographical summary of the number of scientific publications registered in the arXiv (https://arxiv.org/) and Science Direct (https://www.sciencedirect.com/) databases presenting the terms “machine learning” (ML) and “deep learning” (DL) in relation with different types of skin cancer (consulted 1 July 2021). Searches considered the appearance of these terms in either the abstract, title, or keywords.
Table 1. A bibliographical summary of the number of scientific publications registered in the arXiv (https://arxiv.org/) and Science Direct (https://www.sciencedirect.com/) databases presenting the terms “machine learning” (ML) and “deep learning” (DL) in relation with different types of skin cancer (consulted 1 July 2021). Searches considered the appearance of these terms in either the abstract, title, or keywords.
arXivScience Direct
MLDLMLDL
Skin Cancer47564371
Non-Melanoma Skin Cancer (NMSC)2355
Melanoma13157578
Cutaneous Squamous Cell Carcinoma (SCC/cSCC)3352
Basal Cell Carcinoma (BCC)6359
Table 2. Description of the final model architecture used for the supervised classification of hyperspectral signatures. The 1D Inception module blocks are constructed following the architecture presented in Figure 2.
Table 2. Description of the final model architecture used for the supervised classification of hyperspectral signatures. The 1D Inception module blocks are constructed following the architecture presented in Figure 2.
Convolutional Neural Support Vector Machine
Input: 1 × 94 Vector Hyperspectral Signature
1D Inception Module
Concatenation
1D Inception Module
Concatenation
Flattening
Dropoutp = 0.54
Dense = 200
Dropoutp = 0.33
Dense = 150
Dropoutp = 0.10
Dense = 100
Dropoutp = 0.46
Dense = 50
Radial Kernel Support Vector Machine Activation
Binary Output label: Healthy (0) or BCC (1)
Table 3. Algorithm performance on test sets. AUC = area under curve. MSE = mean squared error. ReLU = rectified linear unit. SGD = stochastic gradient descent.
Table 3. Algorithm performance on test sets. AUC = area under curve. MSE = mean squared error. ReLU = rectified linear unit. SGD = stochastic gradient descent.
Swish and AdamReLU and AdamSwish and SGDReLU and SGD
Accuracy0.900.820.900.91
Sensitivity0.850.710.890.89
Specificity0.940.930.920.92
AUC0.900.820.900.91
Kappa0.790.640.810.81
MSE0.0340.0780.0290.035
Table 4. Model training time (seconds per epoch) using different computer systems as well as specifying the number of CPUs and GPUs made available to Tensorflow during training.
Table 4. Model training time (seconds per epoch) using different computer systems as well as specifying the number of CPUs and GPUs made available to Tensorflow during training.
ComputerNo. CPUsNo. GPUsSeconds/Epoch
Personal Laptop405.94
Desktop Computer404.75
SCAYLE405.36
SCAYLE1002.68
SCAYLE1801.86
SCAYLE410.25
SCAYLE1810.20
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Courtenay, L.A.; González-Aguilera, D.; Lagüela, S.; Pozo, S.D.; Ruiz, C.; Barbero-García, I.; Román-Curto, C.; Cañueto, J.; Santos-Durán, C.; Cardeñoso-Álvarez, M.E.; et al. Deep Convolutional Neural Support Vector Machines for the Classification of Basal Cell Carcinoma Hyperspectral Signatures. J. Clin. Med. 2022, 11, 2315. https://doi.org/10.3390/jcm11092315

AMA Style

Courtenay LA, González-Aguilera D, Lagüela S, Pozo SD, Ruiz C, Barbero-García I, Román-Curto C, Cañueto J, Santos-Durán C, Cardeñoso-Álvarez ME, et al. Deep Convolutional Neural Support Vector Machines for the Classification of Basal Cell Carcinoma Hyperspectral Signatures. Journal of Clinical Medicine. 2022; 11(9):2315. https://doi.org/10.3390/jcm11092315

Chicago/Turabian Style

Courtenay, Lloyd A., Diego González-Aguilera, Susana Lagüela, Susana Del Pozo, Camilo Ruiz, Inés Barbero-García, Concepción Román-Curto, Javier Cañueto, Carlos Santos-Durán, María Esther Cardeñoso-Álvarez, and et al. 2022. "Deep Convolutional Neural Support Vector Machines for the Classification of Basal Cell Carcinoma Hyperspectral Signatures" Journal of Clinical Medicine 11, no. 9: 2315. https://doi.org/10.3390/jcm11092315

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop