Next Article in Journal
Genistein: A Potent Anti-Breast Cancer Agent
Previous Article in Journal
Skeletal Muscle Regeneration by the Exosomes of Adipose Tissue-Derived Mesenchymal Stem Cells
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AoP-LSE: Antioxidant Proteins Classification Using Deep Latent Space Encoding of Sequence Features

1
Department of Computer Engineering, Chosun University, Gwangju 61452, Korea
2
Department of Bio and Brain Engineering, KAIST, Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
Curr. Issues Mol. Biol. 2021, 43(3), 1489-1501; https://doi.org/10.3390/cimb43030105
Submission received: 9 September 2021 / Revised: 28 September 2021 / Accepted: 29 September 2021 / Published: 9 October 2021
(This article belongs to the Section Bioinformatics and Systems Biology)

Abstract

:
It is of utmost importance to develop a computational method for accurate prediction of antioxidants, as they play a vital role in the prevention of several diseases caused by oxidative stress. In this correspondence, we present an effective computational methodology based on the notion of deep latent space encoding. A deep neural network classifier fused with an auto-encoder learns class labels in a pruned latent space. This strategy has eliminated the need to separately develop classifier and the feature selection model, allowing the standalone model to effectively harness discriminating feature space and perform improved predictions. A thorough analytical study has been presented alongwith the PCA/tSNE visualization and PCA-GCNR scores to show the discriminating power of the proposed method. The proposed method showed a high MCC value of 0.43 and a balanced accuracy of 76.2%, which is superior to the existing models. The model has been evaluated on an independent dataset during which it outperformed the contemporary methods by correctly identifying the novel proteins with an accuracy of 95%.

1. Introduction

The oxidative damage due to free radicals is prevented by the antioxidants which are naturally produced in the human body [1]. Small antioxidant molecules, scavenges the free radicals by neutralizing them, where as the large antioxidant molecules prevents the potential damage by absorbing the free radicals [2]. However, on some occasions, the naturally produced antioxidants are outnumbered by the free radicals, and their procurement from the external sources including vegetables and fruits become indispensable [3,4]. The macro-molecules (proteins), that offer these nutrients must be effectively identified, since it would contribute towards the development of novel methods to prevent the diseases caused by the free radicals. Conventionally, the experiments to identify the antioxidant proteins are conducted using biochemical techniques, which however, is arduous and comes at high cost. The protein sequence database provided an opportunity for computer-based antioxidant protein inference [5]. In early study [6], a naive bayes method was developed for the prediction of antioxidant proteins. Amino acid compositions and di-petide compositions of the protein sequences were utilized as features, which were later subjected to a feature selection algorithm. The resultant features with reduced dimension were then used as the training dataset. Later, AoPred was presented by same authors [7], in which g-peptide composition was adopted to extract features from the protein sequences and analysis of variance (ANOVA) [8], was employed to obtain the optimal feature set to train a support vector machines (SVM).
Recently, deep learning (DL) based classification methods have been proposed to perform prediction in almost every field. In particular, many bioinformatics applications have taken advantage of the outstanding capabilities of deep neural networks (DNN) [9]. The prediction of protein structure and function, based on the deep learning algorithms have also been proposed, which yielded highly accurate results [10,11], confirming the potential scope of DL methods in this field. Motivated by the recent success of deep latent-space based learning methods [12,13,14], herein, we propose a deep learning based classification model which is essentially an auto-encoder with an embedded classifier.
The major advantages of the proposed approach is its ability to learn the non-linear embedding of the features whilst maintaining the properties of the original features through reconstruction. This consistency helps filter out fluctuating data points by evaluating their distance from the class centers. As a result, the model can be an important tool for designing a classifier with noisy annotations. Another advantage of using such configuration is the elimination of the need of a crucial step of feature engineering, which involves the selection of dominant features for the training of classifier. The built model has been evaluated by the widely known statistical parameters, and the performance is compared with the existing methods. The results show that the proposed method offers superior performance on the benchmark dataset.
The rest of the paper is organized as follows: In Section 2, the details of the benchmark dataset, features and the classification model are discussed. Section 3 presents the results of the performance evaluation of the proposed method and its comparison with contemporary approaches. The conclusion of the paper is presented in Section 5.

2. Materials and Methods

2.1. Evaluation Parameters

To evaluate the performance of the proposed method, we used metrics for the imbalanced class samples including sensitivity ( S n ) or recall, specificity ( S p ), precision, conventional accuracy ( A C C ), mathews correlation coefficient ( M C C ), balanced accuracy ( B A C C ), youden’s index (YI), F1 score, and Cohen’s kappa ( κ ). These parameters can be evaluated using the following equations:
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
A c c u r a c y = T P + T N T P + T N + F P + F N
M C C = ( T P × T N ) ( F P × F N ) ( T P + F P ) × ( T N + F N ) × ( T P + F N ) × ( T N + F P )
B a l a n c e d A c c u r a c y = S e n s i t i v i t y + S p e c i f i c i t y 2
Y o u d e n s I n d e x = S e n s i t i v i t y + S p e c i f i c i t y 1
F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
P o = T P + T N T P + T N + F P + F N
P e = ( ( T P + F N ) × ( T P + F P ) + ( F P + T N ) × ( F N + T N ) ) ( T P + T N + F P + F N ) 2
κ = P o P e 1 P e
Here T P , F P , T N , and F N represent true positive (correctly classified AoPs), false positive (incorrect classification of non-AoP as AoP), true negative (correctly classified non-AoPs), and false negative (incorrect classification of AoP as non-AoP), respectively.

2.2. Dataset

The benchmark dataset has been obtained from Feng et al. [6]. The formation of the dataset included the selection of the sequences from UniProtKB, that were annotated and reviewed as antioxidant in the molecular function of gene ontology. The sequences were manually observed to remove any nonstandard letters other than the 20 standard amino acid alphabets. Furthermore, the sequence similarity was reduced to ≤60% using CD-HIT. The final dataset contained 253 antioxidant protein sequences and 1552 non-antioxidant proteins. To fairly compare the performance of the proposed method, the training and test datasets were set quantitatively equal to the contemporary approaches, i.e., 200 antioxidant and 1240 non-antioxidant proteins were used as the training subset, while the remaining 53 antioxidants and 312 non-antioxidant proteins were used as the test dataset.

2.3. Features and Latent Space Encoding

The protein features are made compatible with the machine learning algorithms by encoding them numerically. Several encoding schemes have been utilized by the researchers in accordance with the adopted machine learning method [13,15,16,17]. In this study, we use a well known feature encoding method called the composition of k-spaced amino acid pairs (CKSAAP). It is a comprehensive encoding scheme, and has shown significant performance in variety of protein prediction tasks [18,19,20]. In CKSAAP, k denotes the gap or space between the pairs of the amino acid fragments. For k = 2 , the intermediate feature vectors ranging from j = 0 , , k are obtained, which are stacked together to form the final feature vector. An illustration of the feature vector for k = 2 , obtained from the CKSAAP method is shown in Figure 1.
As observed in Figure 1, the dimensions of the resultant feature vector of CKSAAP can be very high. This high resolution feature vector has multilevel granularity which intensifies the relationship of amino acid fragments. However, the inadequate number of positive samples leads to the large p small n problem and therefore, feature engineering such as component analysis [21] and information gain [15] must be employed. The feature engineering tends to work well in the circumstances when the features and class labels have a linear relationship, however, in deep learning applications, this relationship is mostly non-linear, which presents complications in selecting the efficacious features. Therefore, in this study we employ an auto-encoder, which is commonly used for the compressed representation of the input data [22]. The compressed input exists in the bottle-neck of the auto-encoder called the latent space. The latent space contains sufficient information to reconstruct the approximate of the original input by the decoder unit. In the proposed method, the latent space also serves as an input to the classifier module, which is a fully connected neural network, the details of which is presented in Section 2.4.

2.4. Neural Network Architecture

The proposed Deep Latent Space Encoding (DeepLSE) model comprises of three modules: (1) encoder, (2) decoder and (3) classifier. The input layer of the encoder of the baseline model consists of 400 × ( k + 1 ) neurons, the first and second hidden layers has N × 10 , N × 5 neurons respectively, while the final hidden layer has N × 2 neurons. To improve generalization and to avoid over-fitting, the batch-normalization layer and a dropout of 30 % has been employed between each dense layer. The decoder being the complement of the encoder and has mirror symmetric configuration. For classification, a multi-layer perceptron has been implemented with two fully connected hidden layers. Each hidden layer has 10 neurons while the output layer has 2 neurons characterizing the antioxidant and non-antioxidant proteins class labels. For N = 5 , the architecture of the proposed model is shown in Figure 2.

2.5. Training Configurations

The model is trained using Python on Tensorflow-Keras [23] platform for different configurations of latent variables ( L V ) and gap values (k). The output of encoder/decoder and classification network uses sigmoid and soft-max activation functions respectively, while the hidden layers of all modules use rectified linear unit (ReLU) activation function to avoid the occurrence of vanishing gradient. Two loss functions; mean squared error and binary cross-entropy for auto-encoder and classifier respectively are being minimized using the default learning rate of RMSprop optimizer for 1000 epochs with an early stopping tolerance of 100 epochs. The convex combination of two losses is achieved by
L c o m b i n e d = λ L d e c o d e r + ( 1 λ ) L c l a s s i f i c a t i o n ,
where λ is a mixing weight and which was set to be 0.99 .

3. Results and Discussion

3.1. Ablation Study

3.1.1. Finding Best Latent-Space Encoding (LSE) Scheme

The workflow of the proposed study is aimed to obtain the best classification model based on two variable parameters, i.e., the gap between the two amino acid pairs, which is done by setting different values of k during the encoding, and the number of units in the latent space LS. An ablation study has been designed to acquire latent-space encoding (LSE) with varying number of aforementioned variable parameters according to the workflow depicted in Figure 3a. The proteins are distributed into train and test datasets as discussed in the Section 2.2 and resultant datasets are processed for CKSAAP encoding. For the model configuration ( N = 5 ) shown in Figure 2, nine different subsets are constructed by keeping k = 1 to 9. For each value of k, the model is trained with several values of latent space variables L V s ranging from 2 to 10. For each configuration, 20 independent trials are performed and the obtained test results are averaged. Same procedure is followed for the next configuration by incrementing the gap value (k) between the amino acid pairs. The model with the best average results is finally selected as the base model to perform prediction and is named as AoP-LSE.
The summary of the ablation study is provided in a visual and tabular form. The results of the BACC, MCC and PSNR are provided as a surface plot in Figure 4. These metrics are suggested to be an effective evaluation parameter for the imbalanced classification problems. From the results it can be seen that for the large range of parameters, the model achieve comparable accuracy and the variation in both the BACC and MCC is under 10%. However the PSNR performance is highly sensitive to parameter choice as the variation in PSNR around 20 dB units. This clearly indicates discrepancy in most descriptive and most discriminating features. In other words, embedded features which are useful for the reconstruction of original feature space are not necessarily the most discriminating features for designing the classifier, while the most discriminating embedded features are not necessarily the most descriptive (indicative feature for oxidation properties) features. Therefore, finding a balanced features embedding that can help classification while preserving the indicative feature for oxidation properties is important. The same issue is further discussed in the Section 4.1 where learned latent-spaces are compared.
Table 1 presents the results of the balanced accuracy, which is the average of recall of each class and is suggested to be an effective evaluation parameter for the imbalanced classification problems. The consistent values of standard deviations and mean for 20 random trials manifests the stability of the model. However, the best average balanced accuracy results (i.e., 76.2 % ) was obtained when the value of both k and L V was equal to 6. Therefore, the model with k = 6 and L V = 6 has been selected for further testing and evaluations.

3.1.2. Finding Best-Configuration for DeepLSE Architecture

For the aforementioned combination of k = 6 , and L V = 6 , three configurations of model parameters were evaluated. The number of neurons in each configuration were incremented as shown in Table 2, and their effect on the classification performance were evaluated for the metrics including Youden’s index, MCC, receiver operating characteristics area under the curve (ROC AUC), precision recall area under the curve (PR AUC) and the mean reconstruction error. 10 independent trials were performed for each configuration and the one producing the best results in terms of aforementioned performance metrics was selected as the baseline model in this study. Table 3 shows the results of the ablation study.
It was observed from the analysis above that a shallow model ( N = 1 ) failed to learn the discriminating features, whereas a highly complex model ( N = 10 ) also showed mediocre performance. Interestingly, the moderate size model with the configuration ( N = 5 ) works best. Although, this ablation study is not exhaustive, nevertheless, it gives sufficient clue that the selected parameters are the optimal choice for best performance.

3.2. Comparison with the Contemporary Methods

Once the best model has been identified from the ablation study, it is selected for further testing and evaluation according to the workflow depicted in Figure 3b. For comparative analysis, the model is compared with the Naive Bayes [6] and AODPred(SVM) [7]. Table 4 presents a comparison of the evaluation metrics of the proposed AoP-LSE including the test dataset Accurary ( A C C ), Sensitivity ( S n ), Specificity ( S p ), and Youden’s Index ( Y I ) with Naive Bayes [6] and AODPred(SVM) [7]. Although the sensitivity of proposed method is relatively lower, the proposed AoP-LSE method achieved 0.14 , and 0.03 units higher Youden’s index value than the Naive Bayes and AODPred(SVM) methods respectively. Since, for a highly skewed test dataset, the higher values of a balanced metric e.g., Youden’s index are much desired than the individual class metric, therefore, the proposed AoP-LSE can be considered as a better classification approach.

3.3. Verification on Independent-Dataset of Antioxidant Proteins

For objective evaluation of the performance of the proposed method, we utilize the reviewed antioxidant proteins from UniProtKB/Swiss-Prot [24]. UniProtKB is a high quality, manually annotated and non-redundant protein sequence database, which brings together experimental results, computed features and scientific conclusions. These sequences were totally independent and were not present in the positive of negative datasets of the proposed study. Comparison was performed among AODPred [7], Vote9 [25] and the proposed AoP-LSE methods. The results are presented in Table 5 show that the proposed method successfully predicts 21 out of 22 independent antioxidant proteins achieving an accuracy of 95.4 % . This superior performance of of AoP-LSE indicates that it can be utilized as a useful tool for the annotation of unknown antioxdants.

4. Analysis of Deep Latent-Space Encoding

4.1. Comparison of Feature and Latent-Space Discrimination Capability

The latent space learned by the proposed method has been visualized in Figure 5. For fair comparison, two feature-encoding/ dimension-reduction methods are utilized. For linear and non-linear embedding, PCA [26] and tSNE [27] methods are used respectively. Both methods were provided with the original feature space and deep encoded latent-space data. From Figure 5 it can be clearly seen that the proposed deep latent space encoding method learns to separate the two classes data which are not distinct in the original feature space as can be seen in Figure 5A. The tSNE results in Figure 5B shows interesting patterns, indicating the possibility of noisy labels, which are not as prominent in original features space. Another important characteristics of proposed method is the fact that even with linear encoding method of PCA, the proposed Deep-LSE encoded features looks linearly separable, which is an added advantage as in many bioinformatics problem the simplicity and explainability is more important than the gain in classification scores.
Since both classification and reconstruction modules are trained simultaneously and the training attention is increased towards minimizing the reconstruction loss. The model has more freedom to represent individual samples with its variability. On the other hand, the attention of learning is towards minimizing the decoder’s loss with a low value of the classification weight. It creates an unfair tug of war between the two objectives and results in a cluster of high dimensional data in a compact distinguishable latent-space.

4.2. Comparison of Proposed DeepLSE and Conventional Auto-Encoder-Based Encoding Schemes

Although, the architecture of the proposed model is similar to auto-encoder (AE) + multi-layer perceptron (MLP) neural network. However, the way it has been configured/trained i.e., combined training of AE+MLP makes it different from the conventional approach where AE is trained separately for dimension reduction and obtain the most representative (low-ranked) features in the latent space (LS) for signal reconstruction. The proposed DeepLSE on the other hand, not only learns the most representative features but also impose a classification constraint on the latent variables. Thus, creating a tug of war between the decoder and the classifier forcing the encoder to generate useful features. To emphasize this point and for a fair comparison with existing techniques, an evaluation on three different models has been presented.
We trained the auto-encoder network with similar configuration ( N = 5 , k = 6 and L V = 6 ) as in DeepLSE and extracted the latent space encoding for the entire dataset. Later the encoder latent-variables were used for the design and evaluation of auto-encoder+multi layer perceptron (AE + MLP), auto-encoder+ support vector machine (AE + SVM) and auto-encoder+naive bayes (AE + NB) methods. The models were evaluated for the several evaluation parameters and the results are reported in Table 6. For AE + MLP model, the same MLP configuration was used as in DeepLSE. For AE+SVM, linear SVM [28,29] with Euclidean-distance-based radial-basis-function [30] kernel, balanced class weights and l2 penalty of C = 1000 is used. For AE + NB method, Gaussian Naive Bayes [31] with prior of ( 0.5 , and 0.5 ) is used. Both the SVM and NB methods were implemented using scikit-learn package [32] on Python 3.
As anticipated, the conventional AE fail to generate useful embedding, hence, the performance of all the aforementioned models is poor. To further highlight this point, we provided the visualization for the AE and DeepLSE latent variables in Figure 6, and compare the discrimination potentiality of the original feature-space of DeepLSE’s and AE’s latent-variables using one-dimensional PCA with absolute-GSSMD [33] method. The GSSMD is derived from generalized contrast-to-noise ratio (GCNR) [34] metric, in which the overlap between two distributions is compared. For an ideal classification, a GCNR score of 1 is obtained, which suggests that the two distributions are distinct with no overlap. While for classification in worst-case, the two distribution must be fully overlapping resulting in the GCNR score of 0. It can be observed in Figure 6a, that the conventional AE, with only dimension reduction constraint, obtained the training and testing MSE of 46.21 dB and 46.37 dB, respectively with (1D-PCA-GCNR score = 0.22 ) and does not distinguishes the two classes. While with a comparable MSE error, the proposed method depicted in Figure 6b, presents superior learning capabilities (1D-PCA-GCNR score = 0.91 ) and maps the AoP and non-AoP in separate regions.

4.3. Analysis of Decoder and Residual Error

Herein, we analyzed the discriminating power of input, decoded and residual signals. This provide another interesting characteristics of our proposed DeepLSE. From the 2D visualization of original feature space X , the decoded-output using proposed DeepLSE X = Dec ( Enc ( X ) ) , and the residual of original-input and decoded output of DeepLSE X X in Figure 7a it can be seen that the original feature space X with 1D-PCA-GCNR score of 0.14 units, is highly overlapping/non-linear and have almost no discriminating capabilities in linear-space. Whereas, the proposed DeepLSE method which acts as optimal-transport (OT) [35,36], and shift the data distributions in such a way that the reconstruction error of the original signal remains comparable to the conventional AE as well as the classification power is greatly improved. Its feature space, shown in Figure 7b, results in 1D-PCA-GCNR score of 0.77 units, which is 0.63 units higher than the original discriminating power. The 2D projection of residual error signal X X in Figure 7c looks identical to the original feature space X , however, the class separability in residual error is dropped by 0.02 units, This is understandable as the ideal residual error should not contain any useful information which may improve the classification.

5. Conclusions

In this study, we proposed a deep latent space encoding method for the classification of anti-oxidant proteins using sequence derived features. In particular, the composition of k-spaced amino acid pair (CKSAAP) and a densely connected multi-layered perception neural network are used which are trained in tandem. The proposed method can be used to extract a non-redundant compact feature space, which is shown to outperform the conventional antioxidation protein classification approaches. Furthermore, the effect of varying parameters and number of gaps in CKSAAP and latent variables in Deep-LSE is analyzed, which suggests that a sufficient separable encoding can be learned by keeping moderate number of neurons in the architecture with k = 6 and L V = 6 . The proposed approach effectively learns the non-linear embedding of the features from original feature space, thus filters out the non-relevant data on the basis of their distance from the class centers. This property leads towards the development of a classifier for the antioxidant proteins with noisy annotations and for assessment, the proposed method has been evaluated on an independent dataset which showed superior classification performance compared to the contemporary methods. We hope that AoP-LSE will serve as an effective method for the identification of unknown antioxidants.

Author Contributions

Conceptualization, M.U. and S.K.; methodology, S.K.; software, S.K. and S.P.; validation, S.K. and S.P.; writing—original draft preparation, M.U.; writing—review and editing, M.U. and S.K.; visualization, M.U., S.K. and S.P.; supervision, J.-A.L.; project administration, J.-A.L.; funding acquisition, J.-A.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by research fund from Chosun University (K949856043-1).

Conflicts of Interest

The authors declare that they have no competing interest.

References

  1. Chauvin, J.P.R.; Griesser, M.; Pratt, D.A. The antioxidant activity of polysulfides: It’s radical! Chem. Sci. 2019, 10, 4999–5010. [Google Scholar] [CrossRef] [Green Version]
  2. Sannasimuthu, A.; Arockiaraj, J. Intracellular free radical scavenging activity and protective role of mammalian cells by antioxidant peptide from thioredoxin disulfide reductase of Arthrospira platensis. J. Funct. Foods 2019, 61, 103513. [Google Scholar] [CrossRef]
  3. Tang, J.; Fu, J.; Wang, Y.; Luo, Y.; Yang, Q.; Li, B.; Tu, G.; Hong, J.; Cui, X.; Chen, Y.; et al. Simultaneous improvement in the precision, accuracy, and robustness of label-free proteome quantification by optimizing data manipulation chains. Mol. Cell. Proteom. 2019, 18, 1683–1699. [Google Scholar] [CrossRef] [PubMed]
  4. Grzesik, M.; Bartosz, G.; Stefaniuk, I.; Pichla, M.; Namieśnik, J.; Sadowska-Bartosz, I. Dietary antioxidants as a source of hydrogen peroxide. Food Chem. 2019, 278, 692–699. [Google Scholar] [CrossRef]
  5. Feng, P.; Ding, H.; Lin, H.; Chen, W. AOD: The antioxidant protein database. Sci. Rep. 2017, 7, 1–4. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Feng, P.M.; Lin, H.; Chen, W. Identification of antioxidants from sequence information using naive Bayes. Comput. Math. Methods Med. 2013, 2013, 567529. [Google Scholar] [CrossRef] [Green Version]
  7. Feng, P.; Chen, W.; Lin, H. Identifying antioxidant proteins by using optimal dipeptide compositions. Interdiscip. Sci. Comput. Life Sci. 2016, 8, 186–191. [Google Scholar] [CrossRef]
  8. St, L.; Wold, S. Analysis of variance (ANOVA). Chemom. Intell. Lab. Syst. 1989, 6, 259–272. [Google Scholar]
  9. Li, H.; Tian, S.; Li, Y.; Fang, Q.; Tan, R.; Pan, Y.; Huang, C.; Xu, Y.; Gao, X. Modern deep learning in bioinformatics. J. Mol. Cell Biol. 2020, 12, 823–827. [Google Scholar] [CrossRef]
  10. Senior, A.W.; Evans, R.; Jumper, J.; Kirkpatrick, J.; Sifre, L.; Green, T.; Qin, C.; Žídek, A.; Nelson, A.W.; Bridgland, A.; et al. Improved protein structure prediction using potentials from deep learning. Nature 2020, 577, 706–710. [Google Scholar] [CrossRef]
  11. Torrisi, M.; Pollastri, G.; Le, Q. Deep learning methods in protein structure prediction. Comput. Struct. Biotechnol. J. 2020, 18, 1301–1310. [Google Scholar] [CrossRef]
  12. Park, S.; Khan, S.; Wahab, A. E3-targetPred: Prediction of E3-Target Proteins Using Deep Latent Space Encoding. arXiv 2020, arXiv:2007.12073. [Google Scholar]
  13. Usman, M.; Khan, S.; Lee, J.A. Afp-LSe: Antifreeze proteins prediction Using Latent Space encoding of composition of k-Spaced Amino Acid pairs. Sci. Rep. 2020, 10, 1–13. [Google Scholar] [CrossRef] [PubMed]
  14. Al-Saggaf, U.M.; Usman, M.; Naseem, I.; Moinuddin, M.; Jiman, A.A.; Alsaggaf, M.U.; Alshoubaki, H.K.; Khan, S. ECM-LSE: Prediction of Extracellular Matrix Proteins using Deep Latent Space Encoding of k-Spaced Amino Acid Pairs. Front. Bioeng. Biotechnol. 2021. [Google Scholar] [CrossRef]
  15. Khan, S.; Naseem, I.; Togneri, R.; Bennamoun, M. Rafp-pred: Robust prediction of antifreeze proteins using localized analysis of n-peptide compositions. IEEE/ACM Trans. Comput. Biol. Bioinform. 2016, 15, 244–250. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Naseem, I.; Khan, S.; Togneri, R.; Bennamoun, M. ECMSRC: A sparse learning approach for the prediction of extracellular matrix proteins. Curr. Bioinform. 2017, 12, 361–368. [Google Scholar] [CrossRef]
  17. Usman, M.; Khan, S.; Park, S.; Wahab, A. AFP-SRC: Identification of Antifreeze Proteins Using Sparse Representation Classifier. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  18. Mosharaf, M.P.; Hassan, M.M.; Ahmed, F.F.; Khatun, M.S.; Moni, M.A.; Mollah, M.N.H. Computational prediction of protein ubiquitination sites mapping on Arabidopsis thaliana. Comput. Biol. Chem. 2020, 85, 107238. [Google Scholar] [CrossRef]
  19. Usman, M.; Lee, J.A. Afp-cksaap: Prediction of antifreeze proteins using composition of k-spaced amino acid pairs with deep neural network. In Proceedings of the 2019 IEEE 19th International Conference on Bioinformatics and Bioengineering (BIBE), Athens, Greece, 28–30 October 2019; pp. 38–43. [Google Scholar]
  20. Ju, Z.; Wang, S.Y. Prediction of lysine formylation sites using the composition of k-spaced amino acid pairs via Chou’s 5-steps rule and general pseudo components. Genomics 2020, 112, 859–866. [Google Scholar] [CrossRef]
  21. Zhao, H.; Zheng, J.; Xu, J.; Deng, W. Fault diagnosis method based on principal component analysis and broad learning system. IEEE Access 2019, 7, 99263–99272. [Google Scholar] [CrossRef]
  22. Yoon, Y.H.; Khan, S.; Huh, J.; Ye, J.C. Efficient b-mode ultrasound image reconstruction from sub-sampled rf data using deep learning. IEEE Trans. Med. Imaging 2018, 38, 325–336. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 29 September 2021).
  24. Consortium, U. UniProt: A worldwide hub of protein knowledge. Nucleic Acids Res. 2019, 47, D506–D515. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Li, X.; Tang, Q.; Tang, H.; Chen, W. Identifying antioxidant proteins by combining multiple methods. Front. Bioeng. Biotechnol. 2020, 8, 858. [Google Scholar] [CrossRef]
  26. Jolliffe, I.T. Principal components in regression analysis. In Principal Component Analysis; Springer: Berlin/Heidelberg, Germany, 1986; pp. 129–155. [Google Scholar]
  27. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  28. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011, 2, 1–27. [Google Scholar] [CrossRef]
  29. Fan, R.E.; Chang, K.W.; Hsieh, C.J.; Wang, X.R.; Lin, C.J. LIBLINEAR: A library for large linear classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
  30. Khan, S.; Naseem, I.; Togneri, R.; Bennamoun, M. A novel adaptive kernel for the rbf neural networks. Circuits Syst. Signal Process. 2017, 36, 1639–1653. [Google Scholar] [CrossRef] [Green Version]
  31. Rennie, J.D.; Shih, L.; Teevan, J.; Karger, D.R. Tackling the poor assumptions of naive bayes text classifiers. In Proceedings of the 20th International Conference on Machine Learning (ICML-03), Washington, DC, USA, 21–24 August 2003; pp. 616–623. [Google Scholar]
  32. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  33. Park, S.; Khan, S.; Moinuddin, M.; Al-Saggaf, U.M. GSSMD: A new standardized effect size measure to improve robustness and interpretability in biological applications. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 1096–1099. [Google Scholar]
  34. Rodriguez-Molares, A.; Rindal, O.M.H.; D’hooge, J.; Måsøy, S.E.; Austeng, A.; Bell, M.A.L.; Torp, H. The generalized contrast-to-noise ratio: A formal definition for lesion detectability. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2019, 67, 745–759. [Google Scholar] [CrossRef] [Green Version]
  35. Peyré, G.; Cuturi, M. Computational optimal transport: With applications to data science. Found. Trends Mach. Learn. 2019, 11, 355–607. [Google Scholar] [CrossRef]
  36. Khan, S.; Huh, J.; Ye, J.C. Variational Formulation of Unsupervised Deep Learning for Ultrasound Image Artifact Removal. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2021, 68, 2086–2100. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of CKSAAP descriptor calculation for k = 2 . Extracted from [13].
Figure 1. Illustration of CKSAAP descriptor calculation for k = 2 . Extracted from [13].
Cimb 43 00105 g001
Figure 2. Proposed latent-space encoding-based Antioxidant protein prediction model.
Figure 2. Proposed latent-space encoding-based Antioxidant protein prediction model.
Cimb 43 00105 g002
Figure 3. Workflow and ablation study diagram for AoP-LSE method.
Figure 3. Workflow and ablation study diagram for AoP-LSE method.
Cimb 43 00105 g003
Figure 4. Ablation study: Number of latent variables (LV) and Gaps (k) in CKSAAP features are optimized within 4 × 9 parameter space. (A) Balanced Accuracy, (B) MCC and (C) PSNR.
Figure 4. Ablation study: Number of latent variables (LV) and Gaps (k) in CKSAAP features are optimized within 4 × 9 parameter space. (A) Balanced Accuracy, (B) MCC and (C) PSNR.
Cimb 43 00105 g004
Figure 5. Visualization of Feature and Latent Space Embedding by PCA & tSNE. (A) Feature Space Embeddings (PCA, tSNE), and (B) Latent Space Embeddings (PCA, tSNE). The opaque points are the training instances and clearly expressed points are test instances.
Figure 5. Visualization of Feature and Latent Space Embedding by PCA & tSNE. (A) Feature Space Embeddings (PCA, tSNE), and (B) Latent Space Embeddings (PCA, tSNE). The opaque points are the training instances and clearly expressed points are test instances.
Cimb 43 00105 g005
Figure 6. 2D visualization of (a) latent-space embedding of conventional auto-encoder, and (b) latent-space embedding of proposed DeepLSE-based Auto-Encoder neural network using PCA. (a) Training/Test MSE (dB): 46.21 / 46.37 , and 1D-PCA-GCNR score 0.22 , (b) Training/Test MSE (dB): 45.05 / 44.88 and 1D-PCA-GCNR score 0.91 .
Figure 6. 2D visualization of (a) latent-space embedding of conventional auto-encoder, and (b) latent-space embedding of proposed DeepLSE-based Auto-Encoder neural network using PCA. (a) Training/Test MSE (dB): 46.21 / 46.37 , and 1D-PCA-GCNR score 0.22 , (b) Training/Test MSE (dB): 45.05 / 44.88 and 1D-PCA-GCNR score 0.91 .
Cimb 43 00105 g006
Figure 7. 2D visualization of (a) original feature space (1D-PCA-GCNR score 0.14 ), (b) decoded-output using proposed DeepLSE(1D-PCA-GCNR score 0.77 ) and (c) residual of original-input and decoded output of DeepLSE neural network using PCA (1D-PCA-GCNR score 0.12 ).
Figure 7. 2D visualization of (a) original feature space (1D-PCA-GCNR score 0.14 ), (b) decoded-output using proposed DeepLSE(1D-PCA-GCNR score 0.77 ) and (c) residual of original-input and decoded output of DeepLSE neural network using PCA (1D-PCA-GCNR score 0.12 ).
Cimb 43 00105 g007
Table 1. Balanced accuracy results of ablation study on Gap(k), and L V parameters.
Table 1. Balanced accuracy results of ablation study on Gap(k), and L V parameters.
Gap/LV23456789
k = 1 0.739 ± 0.082 0.713 ± 0.070 0.716 ± 0.091 0.688 ± 0.094 0.710 ± 0.047 0.717 ± 0.139 0.726 ± 0.049 0.704 ± 0.109
k = 2 0.701 ± 0.080 0.725 ± 0.068 0.715 ± 0.086 0.675 ± 0.079 0.717 ± 0.057 0.748 ± 0.026 0.641 ± 0.124 0.654 ± 0.093
k = 3 0.729 ± 0.083 0.759 ± 0.034 0.702 ± 0.110 0.716 ± 0.061 0.668 ± 0.124 0.703 ± 0.083 0.722 ± 0.094 0.669 ± 0.117
k = 4 0.722 ± 0.080 0.648 ± 0.107 0.702 ± 0.082 0.708 ± 0.089 0.685 ± 0.111 0.670 ± 0.108 0.678 ± 0.116 0.752 ± 0.051
k = 5 0.721 ± 0.036 0.670 ± 0.120 0.688 ± 0.105 0.707 ± 0.096 0.680 ± 0.090 0.708 ± 0.078 0.703 ± 0.073 0.702 ± 0.064
k = 6 0.689 ± 0.114 0.702 ± 0.087 0.692 ± 0.085 0.717 ± 0.077 0 . 762 ± 0 . 062 0.739 ± 0.061 0.740 ± 0.096 0.753 ± 0.019
k = 7 0.699 ± 0.078 0.721 ± 0.079 0.678 ± 0.066 0.713 ± 0.044 0.703 ± 0.098 0.713 ± 0.075 0.657 ± 0.095 0.710 ± 0.098
k = 8 0.720 ± 0.062 0.711 ± 0.069 0.636 ± 0.128 0.694 ± 0.104 0.676 ± 0.101 0.681 ± 0.106 0.725 ± 0.034 0.703 ± 0.098
k = 9 0.718 ± 0.065 0.708 ± 0.086 0.689 ± 0.066 0.729 ± 0.085 0.694 ± 0.098 0.734 ± 0.063 0.694 ± 0.120 0.720 ± 0.045
Table 2. Number of neurons in hidden layers of the architecture for the tested configurations.
Table 2. Number of neurons in hidden layers of the architecture for the tested configurations.
Configuration (N)1510
Encode10-5-250-25-10100-50-20
Decode2-5-1010-25-5020-50-100
Classifier2-210-1020-20
Table 3. Training and test results of ablation study performed on network parameters.
Table 3. Training and test results of ablation study performed on network parameters.
Metric/Configuration (N)1510
TrainTestTrainTestTrainTest
Youden’s Index 0.53 ± 0.31 0.35 ± 0.20 0.95 ± 0.08 0.59 ± 0.06 0.86 ± 0.08 0.52 ± 0.07
MCC 0.35 ± 0.20 0.45 ± 0.31 0.83 ± 0.15 0.47 ± 0.06 0.73 ± 0.14 0.42 ± 0.06
ROC-AUC 0.80 ± 0.16 0.70 ± 0.12 0.99 ± 0.01 0.84 ± 0.03 0.97 ± 0.02 0.81 ± 0.03
PR-AUC 0.60 ± 0.21 0.46 ± 0.12 0.95 ± 0.05 0.60 ± 0.07 0.84 ± 0.14 0.52 ± 0.11
MSE ( d B ) 40.81 ± 6.63 42.90 ± 5.27 39.92 ± 4.49 39.82 ± 4.65 35.13 ± 4.47 36.70 ± 4.73
Table 4. Performance statistics of ( k = 6 ,   L V = 6 ) model.
Table 4. Performance statistics of ( k = 6 ,   L V = 6 ) model.
MethodAccuracy S n S p PrecisionYIBACCMCCF1 κ
Naive Bayes [6]0.6680.7200.6600.260.380.6900.270.380.22
AODPred(SVM) [7]0.7470.7500.7440.330.490.7470.360.460.32
AoP-LSE (DL)0.8240.6740.8490.430.520.7620.430.520.42
Table 5. Prediction results for 22 independent antioxidant proteins. “✓” indicates correct identification while “✗” represents a incorrect identification.
Table 5. Prediction results for 22 independent antioxidant proteins. “✓” indicates correct identification while “✗” represents a incorrect identification.
UniProtKB ACCNCBI DefinitionAODPredVote9AoP-LSE
P9WQB7Alkyl hydroperoxide reductase C
P9WHH9Dihydrolipoyl dehydrogenase
P9WIS7Dihydrolipoyllysine-residue
P9WG35Thiol peroxidase
P9WGE9Superoxide dismutase
P9WQB5Alkyl hydroperoxide reductase
P9WIE3Alkyl hydroperoxide reductase
P0CU34Peroxiredoxin TSA1
Q5ACV9Cell surface superoxide dismutase
P9WHH8Dihydrolipoyl dehydrogenase
P9WIE1Putative peroxiredoxin Rv2521
P9WIS6Dihydrolipoyllysine-residue
P9WQB6Alkyl hydroperoxide reductase
P9WID9Putative peroxiredoxin Rv1608c
O17433Cys peroxiredoxin
P9WIE0Putative peroxiredoxin MT2597
P9WID8Putative peroxiredoxin MT1643
P9WGE8Superoxide dismutase [Cu-Zn]
C0HK70Superoxide dismutase
P9WQB4Alkyl hydroperoxide reductase AhpD
P9WG34Thiol peroxidase
P9WIE2Alkyl hydroperoxide reductase E
Table 6. Evaluation of different models trained on the latent variables of an auto-encoder on complete dataset.
Table 6. Evaluation of different models trained on the latent variables of an auto-encoder on complete dataset.
MethodSensitivitySpecificityAccuracyBACCMCCF1 ScoreYI
AE + MLP0.650.570.580.610.160.310.23
AE + SVM0.640.560.570.500.150.300.21
AE + NB0.780.360.420.570.110.280.15
Proposed DeepLSE0.670.840.820.760.430.520.52
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Usman, M.; Khan, S.; Park, S.; Lee, J.-A. AoP-LSE: Antioxidant Proteins Classification Using Deep Latent Space Encoding of Sequence Features. Curr. Issues Mol. Biol. 2021, 43, 1489-1501. https://doi.org/10.3390/cimb43030105

AMA Style

Usman M, Khan S, Park S, Lee J-A. AoP-LSE: Antioxidant Proteins Classification Using Deep Latent Space Encoding of Sequence Features. Current Issues in Molecular Biology. 2021; 43(3):1489-1501. https://doi.org/10.3390/cimb43030105

Chicago/Turabian Style

Usman, Muhammad, Shujaat Khan, Seongyong Park, and Jeong-A Lee. 2021. "AoP-LSE: Antioxidant Proteins Classification Using Deep Latent Space Encoding of Sequence Features" Current Issues in Molecular Biology 43, no. 3: 1489-1501. https://doi.org/10.3390/cimb43030105

Article Metrics

Back to TopTop