Next Article in Journal
Optimal Training Configurations of a CNN-LSTM-Based Tracker for a Fall Frame Detection System
Next Article in Special Issue
Short-Time Impedance Spectroscopy Using a Mode-Switching Nonsinusoidal Oscillator: Applicability to Biological Tissues and Continuous Measurement
Previous Article in Journal
Sensor Size Effect on Rayleigh Wave Velocity on Cementitious Surfaces
Previous Article in Special Issue
Infection-Mediated Clinical Biomarkers for a COVID-19 Electrical Biosensing Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Informed Attentive Predictors: A Generalisable Architecture for Prior Knowledge-Based Assisted Diagnosis of Cancers

School of Informatics, Xiamen University, Xiamen 361001, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(19), 6484; https://doi.org/10.3390/s21196484
Submission received: 27 August 2021 / Revised: 20 September 2021 / Accepted: 24 September 2021 / Published: 28 September 2021
(This article belongs to the Special Issue Advanced Biosensors for Human Disease Detection and Monitoring)

Abstract

:
Due to the high mortality of many cancers and their related diseases, the prediction and prognosis techniques of cancers are being extensively studied to assist doctors in making diagnoses. Many machine-learning-based cancer predictors have been put forward, but many of them have failed to become widely utilised due to some crucial problems. For example, most methods require too much training data, which is not always applicable to institutes, and the complicated genetic mutual effects of cancers are generally ignored in many proposed methods. Moreover, a majority of these assist models are actually not safe to use, as they are generally built on black-box machine learners that lack references from related field knowledge. We observe that few machine-learning-based cancer predictors are capable of employing prior knowledge (PrK) to mitigate these issues. Therefore, in this paper, we propose a generalisable informed machine learning architecture named the Informed Attentive Predictor (IAP) to make PrK available to the predictor’s decision-making phases and apply it to the field of cancer prediction. Specifically, we make several implementations of the IAP and evaluate its performance on six TCGA datasets to demonstrate the effectiveness of our architecture as an assist system framework for actual clinical usage. The experimental results show a noticeable improvement in IAP models on accuracies, f1-scores and recall rates compared to their non-IAP counterparts (i.e., basic predictors).

1. Introduction

Cancer is a major cause of human death, as it often leads to a series of related fatal diseases that are complicated to control [1]. As a common sign of cancer, cancer tumours are often caused by different genetic mutations and generally display considerable phenotypic heterogeneity in cancer cells [2]. It is, therefore, a common challenge among researchers to eliminate cancers, because the genetic pathogenic mechanisms of cancer tumours are too complex to be observed or properly understood [3].
Due to its complexity, the clinical treatment of cancer is in high demand, the goal of which is more practical and real-life reasonable. The early detection and prediction techniques of cancers, which facilitate successive medical treatment of patients, are being extensively studied [4] in order to assist doctors in making early diagnoses, thus improving the survival rate of cancer cases and helping discover useful insights into disease pathogenesis [5].
With the improvement of device performance and accuracy, DNA sequencing techniques are now able to collect genomic profiles of tumour samples, such as Gene Expression (GE), miRNA Expression (ME) and DNA Methylation (DM), without extensive physical or mental exertion. While it is demanding to excavate possible patterns of tumour-causing gene expression signatures from clinical data for even well-experienced physicians, many data-driven approaches can be used to assist doctors in the early detection of cancers. Because of their ability to discover feature patterns from large-scale data, machine learning algorithms such as Support Vector Machines (SVMs), Bayesian Networks (BNs) and Decision Trees (DTs) have been well employed for modelling the development and treatment of cancer conditions [6,7,8]. These methods are useful for early cancer detection, as they can automatically capture the effective patterns of samples, which are more likely to be diagnosed as tumour tissue.
Many machine learning methods use gene expression signatures as inputs, which indicates a distinctive, species- or tissue-specific expression pattern of genes in one cell [9]. While it is true that gene signatures could considerably enhance the capabilities of cancer prognosis, many relevant methods are not suitable for medical practice, as they lack necessary intelligent frameworks when facing the large, complex features of real-world cancer data [10]. Hence, deep learning techniques such as Artificial Neural Networks (ANNs), Recurrent Neural Networks (RNNs) and Deep Neural Networks (DNNs) are integrated to improve the abstractness of cancer prediction and prognosis models, resulting in lower training time, better feature extraction quality and enhanced diagnosis accuracy compared to machine learning techniques [11].
There are a series of deep learning models that achieve elevated performance on certain cancer prediction tasks [12,13,14]. However, the structure of many deep networks is too complex to be understood, as they are often composed of multiple deep layers where high-level, abstract-entity representations are transformed in between. In addition, the gene signature of every sample can vary to a huge extent. A complex, nonlinear and arbitrary decision boundary is often associated with cancer diagnosis [15]. Thus, to mitigate the possibility of overfitting, a large learning dataset is often of necessity for the training process of deep learners [13], which is not necessarily available for certain diseases, and may lead to common data-driven failures such as gradient vanishing [16,17].
Recently, the notion of IML has gained a growing interest, which is especially considered a potential solution for dealing with insufficient training data [18]. Unlike traditional approaches, where knowledge is generally used for preprocessing or feature engineering on training data, IML defines a process through which the model could learn from a composite of multiple information sources that consist of both data and prior knowledge (PrK); the latter is explicitly integrated into the machine learning pipeline from an independent knowledge source [18]. By integrating human experience, e.g., equations, probabilities and knowledge graphs, as PrK into the learning process, the reliability and robustness of the underlying model can be improved [19].
According to our knowledge, few methods in relevant fields are already able to integrate the novel concept of IML into the cancer prediction task, and therefore, few approaches are capable of leveraging PrK to assist cancer diagnosis. Inspired by this, we designed and proposed a conceptual architecture, the Informed Attentive Predictor (IAP), which aims at the cancer prediction task, as well as its extended field application, namely assisted diagnosis of cancers. Moreover, the IAP embodies the idea of informed machine learning (IML) and is designed to be generalisable, i.e., it describes a method of knowledge integration where the human experience straightforwardly takes part in the decision-making process, which is classifier-independent. Several IAP implementations were constructed and evaluated to demonstrate their effectiveness on either full- or reduced-size datasets.
To summarise, our main contributions are threefold:
  • We propose an end-to-end IML architecture (IAP) for the assisted diagnosis of cancers, which describes a method through which the underlying classifier can fuse human experience into its whole process, even if the training data have a limited size. Because the PrK integration module is dependent on the underlying classifier, the IAP is totally generalisable. To the best of our knowledge, the IAP is the first attempt to incorporate IML with cancer prediction.
  • Based on some previously proposed classifiers targeted at the same task, we developed and evaluated several IAP implementations. Specifically, we compared the performance between each IAP implementation and its basic counterpart. Experiments show that by applying the IAP architecture, model performance is improved by a noticeable margin on either full- or reduced-size datasets.
  • The comparative analyses amongst multiple experiments show that our proposed architecture and models achieve higher accuracies, F1-scores and recall on the tested TCGA datasets. Thus, the validity and performance of our methodologies are verified.

2. Related Work

In this section, we exemplify related work targeting the cancer prediction task. Specifically, we focus on methods with machine learning and/or deep learning techniques, as well as some recent applications of IML in relevant fields.

2.1. Machine Learning Methods for Cancer Prediction

Machine Learning (ML) is a branch of Artificial Intelligence (AI) research that utilises a variety of statistical and probabilistic techniques to automatically learn patterns of features from past examples (train) and then use the trained model to classify new data [20].
Traditional Artificial Neural Networks (ANNs) [21], Support Vector Machines (SVMs) [22] and Decision Trees (DTs) [23] have been widely used in cancer detection and diagnosis for approximately two decades [24,25,26]. In [27,28,29], different types of ANNs are employed to build effective cancer prediction models to learn from cell characteristics and gene subsets. The authors of [30,31,32] utilise microarray gene expression data to train their models in cancer prediction and prognosis using ANN algorithms. Refs. [33,34,35,36] presents examples of SVMs used in cancer prediction and prognosis. Apart from methods that employ only one ML technique, many articles, including [37,38], propose synthesis models that employ more than two ML algorithms for cancer prediction.

2.2. Deep Learning Methods for Cancer Prediction

With the development of computational devices, Deep Learning (DL) methods started to play an important role in the construction of classifiers. A DL model has a branch of multiple processing layers, through which representations of data can be learnt in a way that contains different levels of abstraction [39].
A majority of DL methods utilise the Convolution Neural Network (CNN), a feed-forward network that allows for effective feature extraction by applying convolution on embeddings. For instance, in [40], the authors propose a model where multiple CNN-based architectures are implemented and employed to distinguish cancer of different types based on a heterogeneous dataset. In addition, CNN-based models were and are being used in a wide range of cancer prediction tasks, including those that employ Computer Vision (CV) techniques to detect and classify tumour types based on medical images [41,42]. Recently, some newly developed CNN-based DL algorithms have been used in cancer detection and prediction. In [36], a deep feature approach, with a GoogleNet model pretrained on ImageNet used as the feature extractor and a polynomial kernel SVM used as the classifier, was used to identify occult invasive disease in patients diagnosed with Ductal Carcinoma in Situ (DCiS) by core needle biopsy.
However, [43] mentioned that several problems arise from the existing deep learning models, such as overfitting, bad performance with small datasets and the struggle to deal with noisy features. Therefore, several techniques that make use of novel fields of deep learning have been proposed to handle some of these problems, such as BDR-CNN-GCN [44], which incorporates Graph Convolution Network (GCN) and CNN; some newly developed deep neural networks, as discussed by the authors of [45], who use Graph Attention Network (GAT) to identify personalised prognostic markers; and Gated Graph Attention (GGAT) network [46], which uses a gating mechanism to enhance its underlying GAT classifier.

2.3. Relevant Development of Informed Machine Learning

As research into cancers has progressed, multiple large-scale datasets have been made available. However, the heterogeneity of these datasets makes it difficult for researchers to construct cancer predictors that learn from multiple data sources. In addition, DL models are becoming much less explainable, as multiple layers have to be stuck to abstract the representations of numerous features.
As a source of information, PrK has been used to enhance the performance of many ML and DL classifiers. The ML techniques that learn from a synthesis of multiple information sources that consist of data and PrK can be summarised by a recently proposed term, i.e., IML [18]. By employing PrK into the ML pipeline (i.e., training and/or prediction process), fewer data are necessary for learning, and the results of small models can be significantly improved.
IML has been used in a wide range of research fields. In [47], the authors use a modern machine learning technique based on random forests to predict Reynolds stress discrepancies in different flows where data are not available, with existing direct numerical simulation databases used to train discrepancy functions. The authors of [48] propose a methodology for using symbolic knowledge in DL by learning a semantic loss function that bridges between neural output vectors and logical constraints. The authors of [49] propose a constraint-based framework for learning user-specified constraints (e.g., physical laws) and simultaneously using them for supervision. This is useful to mitigate the difficulty of using domain expertise to manually specify constraints.
Furthermore, PrK has also been made use of in many biological and medical fields. In [50], a predictor named immunological Elastic-Net (iEN) is integrated with outside immunological knowledge, resulting in the model demonstrating predictions of higher quality for some clinical outcomes on a dataset of large feature size. In [51], protein–protein interaction knowledge is utilised in their gene selection process. This improves the interpretability of gene signatures compared to pure deep learning models.
As to cancer-related methodologies, the authors of [52] incorporate relevant genetic information extracted from former literature with an SVM classifier that enables the usage of known lung adenocarcinoma genes in its cancer classification task. The accuracy is, therefore, improved due to the more effective method of gene selection.
However, we can find a few additional articles on methods that integrate the concept of IML into the cancer prediction task. Therefore, in this study, we propose the IAP that describes an end-to-end method to incorporate the IML concept with basic classifiers to predict cancers based on gene expression signatures. Unlike the aforementioned PrK-based models, we fuse the knowledge in a different way that aims at enhancing prediction instead of features, which is more generalisable because the underlying classifier can be altered or specifically designed for the targeted task.

3. Methodologies

In this section, we describe the formulations of our architecture and the associated task. We initially describe the cancer prediction task that is considered in this work; then, we describe an introduction to the preprocessing of data. Finally, we put forward the structure of our IAP and give a detailed explanation of the methodologies.

3.1. Problem Formulations

The cancer prediction task considered in this work is essentially the classification of samples, which takes features (i.e., gene expression) of samples as input and gives predictions (i.e., cancer or not) accordingly. The aim of this task is to build an errorless and robust classifier by which a high classification accuracy can, even on new samples, be obtained.

3.2. Terminology

The symbols used in this paper are listed in Table 1. Note that the terms ‘sample’, ‘entity’ and ‘node’ in different contexts refer to the same concept in terms of ‘samples’ in a dataset and ‘entities’ or ‘node’ in graph-structured data.

3.3. Data Preprocessing

We follow up several preprocessing steps to extract PrK from its data source. In addition, the original dataset is standardised into graph-structured data to meet the needs of graph-based IAP models.

3.3.1. Construction of Prior Knowledge

As one of our major contributions, we employ the idea of IML in our model by integrating PrK from another data source. The fusing process varies according to different types of PrK and the supporting task. In this work, we use PrK as an aid for the training process in a 4-step manner: (1) PrK extraction, (2) data reduction, (3) PrK encoding and (4) PrK-based difference amplification. The initial two steps, viz. PrK extraction and data reduction, take place in the preprocessing period. The latter two steps are both taken in the training period. We propose their description in later sections.
1. PrK Extraction
The types of available PrK and the corresponding obtaining methods vary for different research contexts, and they require human effort to design the specific procedure for PrK extraction and encoding. For instance, if the training data are served in forms such as WHI images, it would be convenient to employ pathological image patterns of confirmed cases as PrK to enhance the convoluted representations of images. In this work, based on the type of dataset used (i.e., gene expression rate for tumour and normal samples), it is natural to find relevant gene-related knowledge, be it gene mutation rates of the cancers released by a recognized authority, relative expression levels of different genes recorded in an NIH database, etc.
Through careful investigation, we eventually selected mutation rate as the source of PrK, because it fits well to the main procedure of knowledge integration (Section 3.4.2) and is not model specific. We extracted a statistical table reported by a previous study towards the cooccurrence and exclusiveness of gene mutations amongst several cancer types [53], where the mutation rates of more than a hundred genes are recorded in percentages. Then, we pulled out the values of each cancer type (viz. each column) from the table, with each represented as a knowledge vector k c R N k . They were then put aside for the later steps of training.
2. Data Reduction
Once all k c were obtained, we extracted the expression values corresponding to the reported N k genes from the feature matrix, thereby producing a ‘reduced’ feature vector e ˜ i R N k for each entity i. e ˜ i is formulated in Equation (1)
e i ˜ = e i m 1 , e i m 2 , , e i m N k T
where e i m g ( g 1 , 2 , , N k ) indicates the corresponding feature value (viz. gene expression value) in e i in terms of the g-th reported gene. The ‘reduced’ feature vectors of all samples were produced and saved for later utilisation.
The matrices and vectors produced in the preprocessing period, i.e., adjacency matrix A ( A R N × N ), feature matrix E = e 1 , e 2 , , e N ( E R N e × N ), knowledge vectors and reduced feature vectors, are inputs of our proposed model.

3.3.2. Graph Construction

We standardised the original datasets into undirected and unweighted graphs, as some graph-based classifiers (e.g., GAT, GCN, GGAT) take graph-structured data as their input. For simplicity, we constructed the graph according to the vector distances between the original samples. Specifically, we assumed two samples to be ‘correlated’ if their vector distance equalled or was smaller than the threshold. We used cosine similarity as the metric of the distance between entities, as described in Equation (2).
l i j = e i · e j e i e j = m = 1 N e e i m e j m m = 1 N e e i m 2 m = 1 N e e j m 2
The adjacency matrix of the graph was then constructed according to Equation (3)
A i j = 1 if l i j θ i , 0 otherwise
where θ i is the threshold value for entity i. We used the mean distance of all neighbours as the threshold, i.e.,
θ i = j N i l i j / | N i |

3.4. IAP Structure and Procedure

The IAP architecture can be divided into modules, viz. (1) Underlying Classifier Module (UCM) and (2) Knowledge Integration Module (KIM). They are trained in an end-to-end manner. The overall structure of IAP is shown in Figure 1, where the UCM is rendered as a black box. As an example, Figure 2 depicts the data flow of an IAP implementation of GGAT [46] (IAP-GGAT).
The subsequent part will follow up the division in Figure 1 and give details of the two modules and their submodules accordingly.

3.4.1. Underlying Classifier Module

The UCM aims at giving a basic prediction result based on the original dataset. The prediction vectors are then fed into the KIM and enhanced therein. The enhancements will be formalised in detail in Section 3.4.2. Here, we treat the UCM as a black box that takes feature vectors as input and produces raw prediction vectors ( p i ( raw ) R 2 ). The overall procedure of the UCM can be formalised as a function P ( raw ) = UCM E , where P ( raw ) = p 1 ( raw ) , p 2 ( raw ) , p N ( raw ) indicates the raw predictions set against all samples. The function UCM · indicates the abstract procedure of the UCM. For graph-based UCMs, the adjacency matrix is taken into consideration, resulting in the graph-based UCM overall procedure that is formalised as the function P ( raw ) = UCM E , A .
To be precise, P ( raw ) contains logits of the classifier from which predictions can directly be derived. To improve their quality, the logits are enhanced by an ‘injection’ of PrK during the procedure of KIM.

3.4.2. Knowledge Integration Module

As a key feature of IML, the KIM serves as an extension to the UCM. It learns an effective way to fuse PrK into the prediction process (that is, to take the 3rd and 4th steps of the IAP mentioned in Section 3.3.1). These two steps are performed accordingly by the two key parts of the module: (1) PrK Encoder and (2) PrK Amplifier.
PrK Encoder is used to map both PrK and reduced data into a uniformed vector space and therein produce a separate prediction that depends solely on the PrK. This simple prediction is not intended for practical use. We only utilise the prediction vectors to amplify and enhance the logits from the UCM.
The PrK encoder follows a 2-step procedure: sample generalisation and likelihood calculation. In the ‘sample generalisation’ period, PrK Encoder calculates a general representation for normal samples by applying a linear transform to the ‘reduced’ feature vectors, as indicated by Equation (5)
e ˜ ( norm ) = i = 1 V ( norm ) v i e ˜ V i ( norm ) = E ˜ ( norm ) v
where V ( norm ) indicates a set that consists of the reference numbers of normal samples, and V i ( norm ) is the i-th element of V ( norm ) . E ˜ ( norm ) R N k × V ( norm ) is a reduced feature matrix of normal samples, and v R V ( norm ) is a projection vector that squashes E ˜ ( norm ) into a one-column matrix, i.e., the general representation for normal samples ( e ˜ ( norm ) R N k ).
e ˜ ( norm ) can be seen as a learnt representative of normal samples, as the projection vector v can be optimised as training progresses.
In the ‘likelihood calculation’ period, PrK Encoder measures the exact difference between the reduced feature vector of each sample (regardless of its actual type, i.e., normal or tumour) and the normal sample representative, which is described as Equation (6).
d i = e ˜ ( norm ) e ˜ i e ˜ ( norm )
In d i R N k , each component indicates the relative difference of the corresponding gene expression value with respect to its counterpart in the normal sample representative. Intuitively, d i measures the extent to which i and normal samples are dissimilar in terms of gene expression, and so does the gene mutation rate. Inspired by this, a value of how likely this sample can be categorised as tumour type c can be obtained by calculating an element-wise similarity between d i and k c , as described in Equation (7)
s i c = sim d i , k c = sim e ˜ ( norm ) e ˜ i e ˜ ( norm ) , k c
where the function sim a , b is used to calculate the similarity between two vectors, a and b . Here, considering that the function takes two vectors ( d i and k c ) of different semantic meanings, we employ cosine similarity, which is not sensitive to the exact components values being compared, to mitigate the potential influence of the mismatched magnitudes between them. The output, s i c 0 , 1 , is a scalar, indicating the likelihood of sample i being diagnosed as tumour type c with only a reference to PrK. A prediction vector can thereby be constructed according to Equation (8)
p i ( pk ) = 1 s i c , s i c T
with the first and second components of p i ( pk ) R 2 indicating the prediction of sample i being a normal and tumour sample, respectively. This prediction vector, as previously mentioned, is not intended to make any final prediction or diagnosis. It is only a by-product that will then be further utilised in PrK Amplifier.
PrK Amplifier uses p i ( pk ) to amplify and enhance the results of raw prediction vectors and was previously introduced in Section 3.4.1. The procedure of PrK Amplifier can, likewise, be broken down into 2 simple steps: likelihood weighting and knowledge injection.
In the ‘likelihood weighting’ phase, a weighted version of each p i ( pk ) is obtained, ( p i ( weighted ) ), by assigning different weights to the prediction values. Each p i ( weighted ) is then ‘injected’ into the raw prediction vector accordingly by a simple summation in the ‘knowledge injection’ phase. This action will eventually result in an amplification or weakening of either component in the prediction vector.
The two steps of PrK Amplifier can be jointly formalised as Equation (9)
p i = softmax p i ( raw ) + w p i ( pk )
where w R 2 is optimised as the training progresses. Being normalised by the softmax function, the resulting p i is the ‘adjusted’ prediction vector whose first and second components indicate the final prediction of sample i being a normal and a tumour sample, respectively.
p i is also the output of IAP, on which the optimiser calculates loss and performs fine-tuning on parameters through gradient descending.

3.4.3. Example of IAP Implementation

As an example, Figure 2 depicts the data flow of an IAP implementation of GGAT [46] (IAP-GGAT).
In this example, the domain data, i.e., gene signature of cancer/normal samples, are preprocessed to a graph-like structure. The graph is then fed into the first GGAT layer of the UIM, followed by two subsequent processes performed by the latter two GGAT layers. Furthermore, as in KIM, the reduced feature vectors are squashed into one, which is then used by the encoder to give predictions based on knowledge and features. The prediction vectors produced, i.e., p ( pk ) , are ‘injected’ into the logits of the final GGAT layer, i.e., p ( raw ) , through a procedure that we call ‘knowledge injection’. The injected prediction vectors are the final results of the whole model and will finally be used in calculating the loss function and model optimisation.

4. Evaluation

We implemented four IAP models based on two traditional machine learning classifiers that were previously utilised in similar cancer prediction tasks [35,54], i.e., Support Vector Machine (SVM) and Decision Tree (DT), and two newly developed graph-based deep learning classifiers, i.e., Graph Attention Network (GAT) and Gated Graph Attention Network (GGAT). These models are identified by their original names plus a prefix ‘IAP-’, i.e., IAP-SVM, IAP-DT, IAP-GAT and IAP-GGAT, respectively.
To evaluate the effectiveness of the IAP, we carried out cancer prediction experiments on several TCGA datasets that are available to the public. All of these datasets were initially preprocessed (Section 3.3) into graph-like versions for the four graph-based candidate models (GAT, IAP-GAT, GGAT, IAP-GGAT). Furthermore, we extracted the PrK from [53] and constructed it into specific formats. All experiments were performed in a unified environment, where PyTorch [55] was used to implement our models. In addition to normal experiments, we give details on a supportive experiment that was intended for demonstrating the effectiveness of IAP with the training data size being reduced.

4.1. Introduction and Modifications to Datasets

The datasets on which we carried out experiments were released by The Cancer Genome Atlas (TCGA), a cancer genomics programme of the U.S. National Cancer Institute and the U.S. National Human Genome Research Institute, each containing gene expression values (gene signatures) of samples from either a specific type of tumour or normal tissues. Detailed information on TCGA can be found in https://www.cancer.gov/about-nci/organization/ccg/research/structural-genomics/tcga (accessed on 28 September 2021). The datasets can be downloaded from the Genomic Data Commons (GDC) Data Portal (https://portal.gdc.cancer.gov/ (accessed on 28 September 2021)) of the U.S. National Institutes of Health (NIH).

4.1.1. Statistics and Visualisation of Datasets

In this work, we chose six datasets (tumour types) for evaluation, because they are covered by both TCGA and the original research article of the PrK: Urothelial Bladder Carcinoma (BLCA), Breast-Invasive Carcinoma (BRCA), Head and Neck Squamous Cell Carcinoma (HNSC), Kidney Renal Clear Cell Carcinoma (KIRC), Lung Adenocarcinoma (LUAD) and Lung Squamous Cell Carcinoma (LUSC). Statistics of these datasets are available in Table 2.
Here, we utilise t-SNE [56] to visualise the samples in these datasets and coloured the nodes based on their type (Tumour or Normal).

4.1.2. Data Balancing

We observed an imbalance between the two tumour types in terms of the number of samples, i.e., the number of tumour samples is much more than that of the normal samples in any dataset. This disparity can cause a variety of problems, such as overclassifying the majority group [57] and slow convergence of the neural network [58]. Several solutions can be applied to couple this issue, including data-level methods such as Random Undersampling (RUS), Random Oversampling (ROS) [59], SMOTE [60] and ADASYN [61], as well as algorithm-level methods such as cost-sensitive learning, which assigns penalties into different classes [62].
It is not reasonable to mitigate the number of any sample types, as the datasets used are all small in size (no more than 1215 samples). In this case, to minimise the effect of data imbalance, we only consider methods that increase the number of minority samples. Advanced upsampling techniques such as ADASYN and SMOTE are generative (that is, they produce new samples according to certain statistical rules). However, the generation of these artificial samples does not involve specialised knowledge. They simply use Euclidean distance and random bias to sample new data. Furthermore, according to the t-SNE visualisation of the datasets (Figure 3, Figure 4 and Figure 5), it can be observed that the areas of the two classes slightly overlap with each other. For some datasets such as BLCA, the two classes are actually not spatially divisible. We believe the main reason for this mismatch is the complex co-relation within gene sequences. This means that these generated samples may be misleading because their spatial location is not necessarily related to the actual classes. Therefore, we applied ROS on the minority class (i.e., normal samples), i.e., we randomly duplicated the normal samples to make sure that the two classes were equal in size when training the models (i.e., Normal and Tumour).

4.2. Experiments

In our experiments, the parameters of our model are initialised using Xavier initialization [63] and are then trained using the Adam optimiser [64] with an initial learning rate set to 5 × 10 3 . Graph-based models (GAT, GGAT) are trained with 1500 epochs, whilst for others, the number of epochs is set to 150. We report each candidate model’s classification performance, measured by a number of different metrics, on each dataset.
To mitigate the occurrence of possible exceptions, we employed k-fold cross-validation in our experiments, where the datasets are randomly divided into five parts, and each is then utilised to test the models. Eventually, the average performance of the five test sets is reported. Considering the size of the datasets, we set k = 5 to ensure that the test sets were not too small whilst enabling adequate repetitive experiments.

4.2.1. Training Process

Here, we plot the changes of accuracies and loss values during GGAT’s and IAP-GGAT’s training process in Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 to study the learning process of IAP (first 1500 epochs). As a comparison, we also plot the same figures for both GAT and IAP-GAT in Appendix A on which the same conclusions can be derived.
Noticeably, for IAP-GGAT, the learning curve has a highly remarkable smoothness, whereas there is observable volatility of GGAT when reaching its plateau. Under extreme circumstances, as in KIRC, for example, where a sudden expansion of loss occurs at approximately the 980th epoch, we can hardly observe even marginal changes in either loss or accuracy of IAP-GGAT at the corresponding moment. Furthermore, the time needed for IAP-GGAT to converge (all less than 100 epochs) is visibly shorter than that for GGAT (between 50 and 400 epochs). These differences highlight two strengths of IAP models: (1) less volatile training curve and (2) less time required for model convergence.

4.2.2. Experimental Results

The experimental results in terms of accuracy, recall and F1-score are shown in Table 3, Table 4 and Table 5, where:
  • The best results are in bold texts;
  • A number marked ‘+’ indicates a higher result in the corresponding pair of models;
  • A number marked ‘=’ indicates that the results of corresponding pair of models are equal to each other;
  • The values in the rightmost column are the count of ‘+’s that the corresponding models achieve on the datasets listed.
The experimental results show that all of our IAP implementations, on average, achieved better performance on a majority of datasets compared to their non-IAP counterparts. Only a marginal difference can be observed where the IAP models ranked second. Furthermore, in most cases, the IAP models achieve more ‘+’s than their basic version, with the exception of recall on (IAP-)SVM, where small differences can be discovered in the experimental results on BRCA and KIRC datasets. We believe the occurrence of this is occasional and can be amended by refining the model’s initial settings (e.g., SVM’s kernel type). Nevertheless, the effectiveness of our PrK integration technique is clearly highlighted.
Table 3. Accuracy on all datasets. (†) PrK-ONLY: a naïve model that only facilitates the PrK for prediction.
Table 3. Accuracy on all datasets. (†) PrK-ONLY: a naïve model that only facilitates the PrK for prediction.
ModelsBLCABRCAHNSCKIRCLUADLUSCAvg.+
DT0.98800.99430.99250.97680.99060.9950+0.9895(1+)
IAP-DT0.9940+0.9943=0.9925=0.9953+0.9953+0.99250.9940+(4+)
SVM0.99100.99430.98750.9861+0.99060.99750.9912(1+)
IAP-SVM0.9970+0.9943=0.9925+0.98600.9953+1.0000+0.9934+(5+)
GAT0.95780.83670.98000.90700.93430.99750.9356(0+)
IAP-GAT0.9910+0.9864+0.9950+0.9837+0.9812+1.0000+0.9896+(7+)
GGAT0.85890.60390.99800.99350.99530.99900.9081(0+)
IAP-GGAT1.0000+0.9977+0.9990+0.9935=0.9981+0.9990=0.9979+(5+)
PrK-ONLY†0.48800.51020.47000.52780.52340.47260.4987N/A
Table 4. Recall on all datasets.
Table 4. Recall on all datasets.
ModelsBLCABRCAHNSCKIRCLUADLUSCAvg.+
DT0.97810.9891+0.97980.95920.99860.99320.9830(1+)
IAP-DT0.9867+0.98840.9798=0.9907+0.9986=0.9932=0.9896+(3+)
SVM0.98330.9897+0.99430.9860+0.99540.99550.9907+(3+)
IAP-SVM0.9940+0.98840.9943=0.96950.9954=1.0000+0.9903(2+)
GAT0.94320.83450.97980.92920.90950.99570.9320(0+)
IAP-GAT1.0000+0.9879+0.9923+0.9986+0.9953+1.0000+0.9957+(7+)
GGAT0.89500.73920.99610.98710.99630.99820.9353(0+)
IAP-GGAT1.0000+0.9932+1.0000+0.9987+1.0000+1.0000+0.9987+(7+)
Table 5. F1-score on all datasets.
Table 5. F1-score on all datasets.
ModelsBLCABRCAHNSCKIRCLUADLUSCAvg.+
DT0.98890.9945+0.9927+0.97700.99080.9951+0.9898(3+)
IAP-DT0.9932+0.99420.99160.9953+0.9952+0.99240.9937+(4+)
SVM0.99150.9948+0.98820.9860+0.99090.99770.9915(2+)
IAP-SVM0.9970+0.99420.9917+0.98440.9953+1.0000+0.9938+(5+)
GAT0.95510.84510.97980.90790.93100.99730.9360(0+)
IAP-GAT0.9908+0.9862+0.9950+0.9846+0.9804+1.0000+0.9895+(7+)
GGAT0.89040.59970.99800.99340.99520.99890.9126(0+)
IAP-GGAT1.0000+0.9978+0.9990+0.9935+0.9981+0.9991+0.9979+(7+)
Additionally, for ablation study purposes, we introduce a control group of experiments performed on a naïve classification model that makes decisions based only on PrK (without other underlying models), whose performance in terms of accuracy is shown as ‘PrK-ONLY’ in Table 3. The PrK-ONLY model utilises p ( pk ) for direct classification. According to our aforementioned discussion, the prediction vectors produced by the PrK integration module are only intended to be used to enhance the prediction of the underlying classification model. In this case, the ‘PrK-ONLY’ model’s substandard performance is therefore within anticipation.

4.2.3. Analysis of the Influence of Training Data Size

As discussed in Section 1, large learning datasets are not necessarily available for certain cancer types and may lead to common data-driven failures such as gradient vanishing. However, small datasets can cause insufficient model training, resulting in incomplete models that are not suitable for real-life applications. To mitigate these issues, PrK is often utilised to deal with insufficient learning materials, as the availability of PrK (employed by the KIM) can reduce the underlying classifier’s reliance on training data. We carried out several supporting experiments to verify that a KIM-employed model is not sensitive to a reduced dataset.
We performed a series of additional experiments using the same settings as those in Section 4.2.2, except that we only used HNSC and LUSC datasets, as the model’s performance on them is less diverged (so the resulting statistics could be more accurate), and the scope of candidate models is narrowed down to (IAP-)GAT and (IAP-)GGAT for simplicity. Specifically, we trained each model using two differently sized training sets (20% and 40%). Each model was trained and tested five times with each training set, i.e., each dataset and model pair were trained 10 times in total. We plotted bar charts (Figure 12) for each metric (accuracy, recall and F1-score) to make comparisons and evaluate the performance of the models. The statistical facts (average and standard deviation values) that we used to plot the bar charts are reported in Appendix B.
The charts point out that the IAP-GGAT achieves the highest scores in the three reported metrics. Furthermore, the IAP models (IAP-GGAT, IAP-GAT) all achieve higher scores than the base models (GGAT, GAT) before and after the training data reduction, which reveals the effectiveness of the KIM. The most significant finding in this additional study is that despite tiny effects of reduced datasets can be observed, IAP models can often mitigate the influence, resulting in acceptable scores in certain circumstances where the training data have a limited size.

4.2.4. Analysis of Training Time

We plotted bar charts to compare the training time of the IAP and normal models, as shown in Figure 13 and Figure 14.
From the bar chats, we can observe a marginal difference between the average training time of a normal model and its IAP implementation (shorter than 1 s). Despite the small increase in training time, the IAPs achieve significantly better performance than the non-IAPs. Furthermore, the increments in training time on different datasets and of different models are roughly the same (that is, the time complexity caused by the introduction of IAP is O ( 1 ) ). This means that for even larger datasets, an IAP model could achieve higher performance whilst controlling the increase in its training time within a certain scope.

5. Conclusions and Future Work

In this paper, we proposed a novel IML architecture, IAP, that includes the KIM for cancer prediction. This model fuses multiple sources of data to make use of more knowledge available at hand. Unlike the other PrK-based models, we fused the knowledge in a different way that aims at enhancing prediction instead of features, which is more generalisable because the underlying classifier can be specifically designed and fine-tuned for the targeted task.
As an initial attempt, the PrK that we used in this study is both simple and unimodal, though it actually aided the prediction process and truly improved the performance. In our future study, we plan to integrate multimodal PrK (including medical images, gene description texts, etc.) into the learning process, thereby providing sufficient information to enhance our implemented IAP model’s performance to a greater extent. Moreover, we will attempt to verify the interpretability of our architecture according to certain specifications so as to ensure its safety of usage under actual production circumstances.

Author Contributions

Conceptualization, M.W., L.Q. and H.L.; methodology, L.Q and H.L.; software, H.L.; validation, H.L.; formal analysis, H.L. and M.W.; investigation, L.Q and H.L.; resources, M.W.; data curation, L.Q and H.L.; writing—original draft preparation, H.L. and L.Q.; writing—review and editing, H.L. and M.W.; visualization, H.L.; supervision, M.W.; project administration, M.W. and H.L.; funding acquisition, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fujian Provincial Department of Science and Technology under Grant No. 2019H0001.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the fact that the human-related datasets used in this study were initially collected and provided by TCGA under ethical surveillance from its Ethics, Law and Policy Group. Detailed information concerning TCGA Ethics and Policies can be found in https://www.cancer.gov/about-nci/organization/ccg/research/structural-genomics/tcga/history/policies (accessed on 28 December 2020).

Informed Consent Statement

Patient consent was waived in this study due to the nature of the data source. The human-related datasets used in this study were provided by a third party (i.e., TCGA) and are public to any research programme. As described in TCGA Ethics and Policies (https://www.cancer.gov/about-nci/organization/ccg/research/structural-genomics/tcga/history/policies (accessed on 28 December 2020)), researchers cannot possess any information regarding the participants of the data-collecting experiments. The authors have neither the authority nor responsibility to fetch personal information (e.g., participant names and contact methods) needed for the approval-obtaining process.

Data Availability Statement

The TCGA datasets were downloaded from the NIH GDC database (https://portal.gdc.cancer.gov/ (accessed on 19 September 2021)).

Acknowledgments

This work was supported in part by the Fujian Provincial Department of Science and Technology under Grant No. 2019H0001. Weijun Xu from Xiamen University was involved in part of the data analysis in this research.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. A Record of Training Process (GAT and IAP-GAT)

In addition to Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, we plotted the changes in accuracies and loss values during GAT and IAP-GAT’s training on all six datasets. Likewise, a much slighter volatility can be observed on IAP-GAT compared to its non-IAP counterpart, GAT. Despite the notable differences in volatility between (IAP-)GAT and (IAP-)GGAT when reaching their plateaus, the effect of knowledge integration is highlighted here as well.
Figure A1. Training process of GAT and IAP-GAT on BRCA dataset.
Figure A1. Training process of GAT and IAP-GAT on BRCA dataset.
Sensors 21 06484 g0a1
Figure A2. Training process of GAT and IAP-GAT on HNSC dataset.
Figure A2. Training process of GAT and IAP-GAT on HNSC dataset.
Sensors 21 06484 g0a2
Figure A3. Training process of GAT and IAP-GAT on KIRC dataset.
Figure A3. Training process of GAT and IAP-GAT on KIRC dataset.
Sensors 21 06484 g0a3
Figure A4. Training process of GAT and IAP-GAT on LUAD dataset.
Figure A4. Training process of GAT and IAP-GAT on LUAD dataset.
Sensors 21 06484 g0a4
Figure A5. Training process of GAT and IAP-GAT on LUSC dataset.
Figure A5. Training process of GAT and IAP-GAT on LUSC dataset.
Sensors 21 06484 g0a5

Appendix B. Statistical Information

Here, we report some statistical information of Figure 12 that was used whilst analysing the influence of training sets’ sizes (Section 5). Table A1 and Table A2 show average and standard deviation values of the performance recorded in GAT’s and IAP-GAT’s 10 runs on LUSC and HNSC with 20% and 40% training sets.
Table A1. Average and standard deviation values of the performance recorded in GAT’s and IAP-GAT’s 10 runs on LUSC and HNSC with 20% and 40% training sets.
Table A1. Average and standard deviation values of the performance recorded in GAT’s and IAP-GAT’s 10 runs on LUSC and HNSC with 20% and 40% training sets.
GATIAP-GAT
20% TS40% TS20% TS40% TS
Avg.Std.Avg.Std.Avg.Std.Avg.Std.
Accuracy0.972550.009210.982510.012510.990000.008340.990010.00501
Recall0.974500.011180.983240.012900.989370.009350.987820.01171
F10.972760.008950.982340.012760.990000.008340.990170.00481
Table A2. Average and standard deviation values of the performance recorded in GGAT’s and IAP-GGAT’s 10 runs on LUSC and HNSC with 20% and 40% training sets.
Table A2. Average and standard deviation values of the performance recorded in GGAT’s and IAP-GGAT’s 10 runs on LUSC and HNSC with 20% and 40% training sets.
GATIAP-GAT
20% TS40% TS20% TS40% TS
Avg.Std.Avg.Std.Avg.Std.Avg.Std.
Accuracy0.974510.022220.980520.013320.987500.010840.995010.00001
Recall0.964200.036620.973940.016200.992980.012240.993400.00432
F10.973890.023140.980270.014260.987680.010660.995110.00024

References

  1. Houten, L.; Reilley, A.A. An investigation of the cause of death from cancer. J. Surg. Oncol. 1980, 13, 111–116. [Google Scholar] [CrossRef]
  2. Liang, M.; Li, Z.; Chen, T.; Zeng, J. Integrative Data Analysis of Multi-Platform Cancer Data with a Multimodal Deep Learning Approach. IEEE/ACM Trans. Comput. Biol. Bioinform. 2015, 12, 928–937. [Google Scholar] [CrossRef] [PubMed]
  3. Hu, X.; Zhang, Z. Understanding the Genetic Mechanisms of Cancer Drug Resistance Using Genomic Approaches. Trends Genet. 2016, 32, 127–137. [Google Scholar] [CrossRef] [PubMed]
  4. Ray, S. A Survey on Application of Machine Learning Algorithms in Cancer Prediction and Prognosis. In Data Management, Analytics and Innovation; Springer: Berlin/Heidelberg, Germany, 2021; pp. 349–361. [Google Scholar]
  5. Liu, C.; Huang, Y.; Ozolek, J.A.; Hanna, M.G.; Singh, R.; Rohde, G.K. SetSVM: An approach to set classification in nuclei-based cancer detection. IEEE J. Biomed. Health Inform. 2018, 23, 351–361. [Google Scholar] [CrossRef]
  6. Hemsley, P.A. An outlook on protein S-acylation in plants: What are the next steps? J. Exp. Bot. 2017, 68, 3155–3164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Kretowska, M. Computational Intelligence in Survival Analysis. In Encyclopedia of Business Analytics and Optimization; IGI Global: Hershey, PA, USA, 2014; pp. 491–501. [Google Scholar]
  8. Petalidis, L.P.; Oulas, A.; Backlund, M.; Wayland, M.T.; Liu, L.; Plant, K.; Happerfield, L.; Freeman, T.C.; Poirazi, P.; Collins, V.P. Improved grading and survival prediction of human astrocytic brain tumors by artificial neural network analysis of gene expression microarray data. Mol. Cancer Ther. 2008, 7, 1013–1024. [Google Scholar] [CrossRef] [Green Version]
  9. Karimi, D.; Nir, G.; Fazli, L.; Black, P.C.; Goldenberg, L.; Salcudean, S.E. Deep Learning-Based Gleason grading of prostate cancer from histopathology Images—Role of multiscale decision aggregation and data augmentation. IEEE J. Biomed. Health Informatics 2019, 24, 1413–1426. [Google Scholar] [CrossRef] [PubMed]
  10. Wu, J.; Lian, C.; Ruan, S.; Mazur, T.R.; Mutic, S.; Anastasio, M.A.; Grigsby, P.W.; Vera, P.; Li, H. Treatment Outcome Prediction for Cancer Patients Based on Radiomics and Belief Function Theory. IEEE Trans. Radiat. Plasma Med Sci. 2019, 3, 216–224. [Google Scholar] [CrossRef]
  11. Murthy, N.S.; Bethala, C. Review paper on research direction towards cancer prediction and prognosis using machine learning and deep learning models. J. Ambient. Intell. Humaniz. Comput. 2021, 1–19. [Google Scholar]
  12. Jing, B.; Zhang, T.; Wang, Z.; Jin, Y.; Liu, K.; Qiu, W.; Ke, L.; Sun, Y.; He, C.; Hou, D.; et al. A deep survival analysis method based on ranking. Artif. Intell. Med. 2019, 98, 1–9. [Google Scholar] [CrossRef]
  13. Shimizu, H.; Nakayama, K.I. A 23 gene–based molecular prognostic score precisely predicts overall survival of breast cancer patients. eBioMedicine 2019, 46, 150–159. [Google Scholar] [CrossRef] [Green Version]
  14. Courtiol, P.; Maussion, C.; Moarii, M.; Pronier, E.; Pilcer, S.; Sefta, M.; Manceron, P.; Toldo, S.; Zaslavskiy, M.; Le Stang, N.; et al. Deep learning-based classification of mesothelioma improves prediction of patient outcome. Nat. Med. 2019, 25, 1519–1525. [Google Scholar] [CrossRef] [PubMed]
  15. Shin, H.C. Hybrid clustering and logistic regression for multi-modal brain tumor segmentation. In Proceedings of the Workshops and Challanges in Medical Image Computing and Computer-Assisted Intervention (MICCAI’12), Lima, Peru, 4–8 October 2012. [Google Scholar]
  16. Hu, Y.; Huber, A.; Anumula, J.; Liu, S.C. Overcoming the vanishing gradient problem in plain recurrent networks. arXiv 2018, arXiv:1801.06105. [Google Scholar]
  17. Su, J. GAN-QP: A novel GAN framework without gradient vanishing and lipschitz constraint. arXiv 2018, arXiv:1811.07296. [Google Scholar]
  18. Von Rueden, L.; Mayer, S.; Garcke, J.; Bauckhage, C.; Schuecker, J. Informed machine learning—Towards a taxonomy of explicit integration of knowledge into machine learning. Learning 2019, 18, 19–20. [Google Scholar]
  19. Deng, C.; Ji, X.; Rainey, C.; Zhang, J.; Lu, W. Integrating Machine Learning with Human Knowledge. Iscience 2020, 23, 101656. [Google Scholar] [CrossRef] [PubMed]
  20. Mitchell, T.M. Does machine learning really work? AI Mag. 1997, 18, 11. [Google Scholar]
  21. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  22. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  23. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  24. Maclin, P.S.; Dempsey, J.; Brooks, J.; Rand, J. Using neural networks to diagnose cancer. J. Med Syst. 1991, 15, 11–19. [Google Scholar] [CrossRef]
  25. Simes, R.J. Treatment selection for cancer patients: Application of statistical decision theory to the treatment of advanced ovarian cancer. J. Chronic Dis. 1985, 38, 171–186. [Google Scholar] [CrossRef]
  26. Listgarten, J.; Damaraju, S.; Poulin, B.; Cook, L.; Dufour, J.; Driga, A.; Mackey, J.; Wishart, D.; Greiner, R.; Zanke, B. Predictive models for breast cancer susceptibility from multiple single nucleotide polymorphisms. Clin. Cancer Res. 2004, 10, 2725–2737. [Google Scholar] [CrossRef] [Green Version]
  27. Hu, Y.; Ashenayi, K.; Veltri, R.; O’Dowd, G.; Miller, G.; Hurst, R.; Bonner, R. A comparison of neural network and fuzzy c-means methods in bladder cancer cell classification. In Proceedings of the 1994 IEEE International Conference on Neural Networks (ICNN’94), San Francisco, CA, USA, 28 June–2 July 1994; Volume 6, pp. 3461–3466. [Google Scholar]
  28. Bevilacqua, V.; Mastronardi, G.; Menolascina, F.; Pannarale, P.; Pedone, A. A novel multi-objective genetic algorithm approach to artificial neural network topology optimisation: The breast cancer classification problem. In Proceedings of the 2006 IEEE International Joint Conference on Neural Network, Vancouver, BC, Canada, 16–21 July 2006; pp. 1958–1965. [Google Scholar]
  29. Cho, S.B.; Won, H.H. Cancer classification using ensemble of neural networks with multiple significant gene subsets. Appl. Intell. 2007, 26, 243–250. [Google Scholar] [CrossRef]
  30. Rajeswari, P.; Reena, G.S. Human liver cancer classification using microarray gene expression data. Int. J. Comput. Appl. 2011, 34, 25–37. [Google Scholar]
  31. Sahu, B.; Mishra, D. A Novel Feature Selection Algorithm using Particle Swarm Optimization for Cancer Microarray Data. PRocedia Eng. 2012, 38, 27–31. [Google Scholar] [CrossRef] [Green Version]
  32. Dev, J.; Dash, S.K.; Dash, S.; Swain, M. A classification technique for microarray gene expression data using PSO-FLANN. Int. J. Comput. Sci. Eng. 2012, 4, 1534. [Google Scholar]
  33. Ford, W.; Land, W. A Latent Space Support Vector Machine (LSSVM) Model for Cancer Prognosis. Procedia Comput. Sci. 2014, 36, 470–475. [Google Scholar] [CrossRef] [Green Version]
  34. Exarchos, K.P.; Goletsis, Y.; Fotiadis, D.I. Multiparametric decision support system for the prediction of oral cancer reoccurrence. IEEE Trans. Inf. Technol. Biomed. 2011, 16, 1127–1134. [Google Scholar] [CrossRef] [PubMed]
  35. Akay, M.F. Support vector machines combined with feature selection for breast cancer diagnosis. Expert Syst. Appl. 2009, 36, 3240–3247. [Google Scholar] [CrossRef]
  36. Zhu, Z.; Harowicz, M.; Zhang, J.; Saha, A.; Grimm, L.J.; Hwang, E.S.; Mazurowski, M.A. Deep learning analysis of breast MRIs for prediction of occult invasive disease in ductal carcinoma in situ. Comput. Biol. Med. 2019, 115, 103498. [Google Scholar] [CrossRef] [Green Version]
  37. Ahmad, L.G.; Eshlaghy, A.; Poorebrahimi, A.; Ebrahimi, M.; Razavi, A. Using three machine learning techniques for predicting breast cancer recurrence. J. Health Med. Inform. 2013, 4, 3. [Google Scholar]
  38. Kim, J.; Shin, H. Breast cancer survivability prediction using labeled, unlabeled, and pseudo-labeled patient data. J. Am. Med Inform. Assoc. 2013, 20, 613–618. [Google Scholar] [CrossRef] [Green Version]
  39. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  40. Khosravi, P.; Kazemi, E.; Imielinski, M.; Elemento, O.; Hajirasouliha, I. Deep convolutional neural networks enable discrimination of heterogeneous digital pathology images. eBioMedicine 2018, 27, 317–328. [Google Scholar] [CrossRef] [Green Version]
  41. Gour, M.; Jain, S.; Sunil Kumar, T. Residual learning based CNN for breast cancer histopathological image classification. Int. J. Imag. Syst. Technol. 2020, 30, 621–635. [Google Scholar] [CrossRef]
  42. Duran-Lopez, L.; Dominguez-Morales, J.P.; Conde-Martin, A.F.; Vicente-Diaz, S.; Linares-Barranco, A. PROMETEO: A CNN-based computer-aided diagnosis system for WSI prostate cancer detection. IEEE Access 2020, 8, 128613–128628. [Google Scholar] [CrossRef]
  43. Zhu, W.; Xie, L.; Han, J.; Guo, X. The application of deep learning in cancer prognosis prediction. Cancers 2020, 12, 603. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, Y.D.; Satapathy, S.C.; Guttery, D.S.; Górriz, J.M.; Wang, S.H. Improved breast cancer classification through combining graph convolutional network and convolutional neural network. Inf. Process. Manag. 2021, 58, 102439. [Google Scholar] [CrossRef]
  45. Althubaiti, S.; Kulmanov, M.; Liu, Y.; Gkoutos, G.; Schofield, P.; Hoehndorf, R. DeepMOCCA: A pan-cancer prognostic model identifies personalized prognostic markers through graph attention and multi-omics data integration. bioRxiv 2021. [Google Scholar] [CrossRef]
  46. Qiu, L.; Li, H.; Wang, M.; Wang, X. Gated Graph Attention Network for Cancer Prediction. Sensors 2021, 21, 1938. [Google Scholar] [CrossRef]
  47. Wang, J.X.; Wu, J.L.; Xiao, H. Physics-informed machine learning approach for reconstructing Reynolds stress modeling discrepancies based on DNS data. Phys. Rev. Fluids 2017, 2, 034603. [Google Scholar] [CrossRef] [Green Version]
  48. Xu, J.; Zhang, Z.; Friedman, T.; Liang, Y.; Broeck, G. A semantic loss function for deep learning with symbolic knowledge. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 5502–5511. [Google Scholar]
  49. Ren, H.; Stewart, R.; Song, J.; Kuleshov, V.; Ermon, S. Adversarial constraint learning for structured prediction. arXiv 2018, arXiv:1805.10561. [Google Scholar]
  50. Culos, A.; Tsai, A.S.; Stanley, N.; Becker, M.; Ghaemi, M.S.; McIlwain, D.R.; Fallahzadeh, R.; Tanada, A.; Nassar, H.; Espinosa, C.; et al. Integration of mechanistic immunological knowledge into a machine learning pipeline improves predictions. Nat. Mach. Intell. 2020, 2, 619–628. [Google Scholar] [CrossRef] [PubMed]
  51. Cun, Y.; Fröhlich, H. Prognostic gene signatures for patient stratification in breast cancer-accuracy, stability and interpretability of gene selection approaches using prior knowledge on protein-protein interactions. BMC Bioinform. 2012, 13, 69. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Guan, P.; Huang, D.; He, M.; Zhou, B. Lung cancer gene expression database analysis incorporating prior knowledge with support vector machine-based classification method. J. Exp. Clin. Cancer Res. 2009, 28, 1–7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Kandoth, C.; McLellan, M.D.; Vandin, F.; Ye, K.; Niu, B.; Lu, C.; Xie, M.; Zhang, Q.; McMichael, J.F.; Wyczalkowski, M.A.; et al. Mutational landscape and significance across 12 major cancer types. Nature 2013, 502, 333–339. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Juneja, K.; Rana, C. An improved weighted decision tree approach for breast cancer prediction. Int. J. Inf. Technol. 2020, 12, 797–804. [Google Scholar] [CrossRef]
  55. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  56. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  57. Johnson, J.M.; Khoshgoftaar, T.M. Survey on deep learning with class imbalance. J. Big Data 2019, 6, 1–54. [Google Scholar] [CrossRef]
  58. Anand, R.; Mehrotra, K.G.; Mohan, C.K.; Ranka, S. An improved algorithm for neural network classification of imbalanced training sets. IEEE Trans. Neural Netw. 1993, 4, 962–969. [Google Scholar] [CrossRef] [Green Version]
  59. Van Hulse, J.; Khoshgoftaar, T.M.; Napolitano, A. Experimental perspectives on learning from imbalanced data. In Proceedings of the 24th International Conference on Machine Learning, Corvallis, OR, USA, 20–24 June 2007; pp. 935–942. [Google Scholar]
  60. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  61. He, H.; Yang, B.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks, Hong Kong, China, 1–8 June 2008. [Google Scholar]
  62. Krawczyk, B. Learning from imbalanced data: Open challenges and future directions. Prog. Artif. Intell. 2016, 5, 221–232. [Google Scholar] [CrossRef] [Green Version]
  63. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  64. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Overall structure of IAP.
Figure 1. Overall structure of IAP.
Sensors 21 06484 g001
Figure 2. Data flow of IAP-GGAT.
Figure 2. Data flow of IAP-GGAT.
Sensors 21 06484 g002
Figure 3. t-SNE visualisation of BLCA and BRCA.
Figure 3. t-SNE visualisation of BLCA and BRCA.
Sensors 21 06484 g003
Figure 4. t-SNE visualisation of HNSC and KIRC.
Figure 4. t-SNE visualisation of HNSC and KIRC.
Sensors 21 06484 g004
Figure 5. t-SNE visualisation of LUAD and LUSC.
Figure 5. t-SNE visualisation of LUAD and LUSC.
Sensors 21 06484 g005
Figure 6. Training process of GGAT and IAP-GGAT on BLCA dataset.
Figure 6. Training process of GGAT and IAP-GGAT on BLCA dataset.
Sensors 21 06484 g006
Figure 7. Training process of GGAT and IAP-GGAT on BRCA dataset.
Figure 7. Training process of GGAT and IAP-GGAT on BRCA dataset.
Sensors 21 06484 g007
Figure 8. Training process of GGAT and IAP-GGAT on HNSC dataset.
Figure 8. Training process of GGAT and IAP-GGAT on HNSC dataset.
Sensors 21 06484 g008
Figure 9. Training process of GGAT and IAP-GGAT on KIRC dataset.
Figure 9. Training process of GGAT and IAP-GGAT on KIRC dataset.
Sensors 21 06484 g009
Figure 10. Training process of GGAT and IAP-GGAT on LUAD dataset.
Figure 10. Training process of GGAT and IAP-GGAT on LUAD dataset.
Sensors 21 06484 g010
Figure 11. Training process of GGAT and IAP-GGAT on LUSC dataset.
Figure 11. Training process of GGAT and IAP-GGAT on LUSC dataset.
Sensors 21 06484 g011
Figure 12. Comparisons amongst models trained on original and the reduced training sets in terms of 3 metrics, i.e., accuracy, recall and F1-score (from left to right). The figures show average model performance (in terms of the corresponding metric) on HNSC and LUSC datasets. Black sticks on top of the bars represent 95% confidence intervals.
Figure 12. Comparisons amongst models trained on original and the reduced training sets in terms of 3 metrics, i.e., accuracy, recall and F1-score (from left to right). The figures show average model performance (in terms of the corresponding metric) on HNSC and LUSC datasets. Black sticks on top of the bars represent 95% confidence intervals.
Sensors 21 06484 g012
Figure 13. Comparison of (IAP-)DT’s and (IAP-)SVM’s average training time.
Figure 13. Comparison of (IAP-)DT’s and (IAP-)SVM’s average training time.
Sensors 21 06484 g013
Figure 14. Comparison of (IAP-)GAT’s and (IAP-)GGAT’s average training time.
Figure 14. Comparison of (IAP-)GAT’s and (IAP-)GGAT’s average training time.
Sensors 21 06484 g014
Table 1. List of symbols.
Table 1. List of symbols.
NotationDescription
NThe number of samples
N k The number of genes available in the PrK data source
e i The feature vector (embedding) of sample i
k c The PrK vector of cancer type c
h i ( l ) The hidden state of entity i in the l-th layer
e i m The m-th component in vector e i
l i j The distance between vectors e i and e j
A The adjacency matrix
A i j The component in the i-th row and j-th column of matrix A
E The feature matrix
Table 2. Statistics of datasets. # Tumour and # Normal indicate the number of samples labelled tumour and normal, respectively. # Edges indicates the number of connections (edges) in the graph that was constructed according to vector distance (see Section 3.3.2).
Table 2. Statistics of datasets. # Tumour and # Normal indicate the number of samples labelled tumour and normal, respectively. # Edges indicates the number of connections (edges) in the graph that was constructed according to vector distance (see Section 3.3.2).
Dataset# Tumour# Normal# Edges
BLCA4141949,661
BRCA1102113418,792
HNSC5004478,373
KIRC53872109,593
LUAD5335986,536
LUSC5024979,198
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, H.; Qiu, L.; Wang, M. Informed Attentive Predictors: A Generalisable Architecture for Prior Knowledge-Based Assisted Diagnosis of Cancers. Sensors 2021, 21, 6484. https://doi.org/10.3390/s21196484

AMA Style

Li H, Qiu L, Wang M. Informed Attentive Predictors: A Generalisable Architecture for Prior Knowledge-Based Assisted Diagnosis of Cancers. Sensors. 2021; 21(19):6484. https://doi.org/10.3390/s21196484

Chicago/Turabian Style

Li, Han, Linling Qiu, and Meihong Wang. 2021. "Informed Attentive Predictors: A Generalisable Architecture for Prior Knowledge-Based Assisted Diagnosis of Cancers" Sensors 21, no. 19: 6484. https://doi.org/10.3390/s21196484

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop