Next Article in Journal
Electrical Stimulation for Preventing Skin Injuries in Denervated Gluteal Muscles—Promising Perspectives from a Case Series and Narrative Review
Next Article in Special Issue
Research Progress of Respiratory Disease and Idiopathic Pulmonary Fibrosis Based on Artificial Intelligence
Previous Article in Journal
Chest X-ray in Emergency Radiology: What Artificial Intelligence Applications Are Available?
Previous Article in Special Issue
Comprehensive Review on the Use of Artificial Intelligence in Ophthalmology and Future Research Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Similarity Measure-Based Approach Using RS-fMRI Data for Autism Spectrum Disorder Diagnosis

1
School of Cyberspace Security, Hainan University, Haikou 570228, China
2
School of Computer Science and Technology, Shandong Technology and Business University, Yantai 264005, China
3
School of Computer Science and Technology, Hainan University, Haikou 570228, China
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(2), 218; https://doi.org/10.3390/diagnostics13020218
Submission received: 25 October 2022 / Revised: 24 December 2022 / Accepted: 28 December 2022 / Published: 6 January 2023
(This article belongs to the Special Issue Advanced Artificial Intelligence in Medical Diagnostics and Treatment)

Abstract

:
Autism spectrum disorder (ASD) is a lifelong neurological disease, which seriously reduces the patients’ life quality. Generally, an early diagnosis is beneficial to improve ASD children’s life quality. Current methods based on samples from multiple sites for ASD diagnosis perform poorly in generalization due to the heterogeneity of the data from multiple sites. To address this problem, this paper presents a similarity measure-based approach for ASD diagnosis. Specifically, the few-shot learning strategy is used to measure potential similarities in the RS-fMRI data distributions, and, furthermore, a similarity function for samples from multiple sites is trained to enhance the generalization. On the ABIDE database, the presented approach is compared to some representative methods, such as SVM and random forest, in terms of accuracy, precision, and F1 score. The experimental results show that the experimental indicators of the proposed method are better than those of the comparison methods to varying degrees. For example, the accuracy on the TRINITY site is more than 5% higher than that of the comparison method, which clearly proves that the presented approach achieves a better generalization performance than the compared methods.

1. Introduction

Children with autism spectrum disorder (ASD) present a range of phenotypes, such as social and communication difficulties, restricted interests, and thinking function loss [1]. Typically, autistic children have lifelong mental retardation, and their life quality is often quite low. More unfortunately, ASD cannot be cured so far. Nevertheless, diagnosis and prompt intervention as early as possible play an important role in improving ASD children’s life quality. Neuroimaging has significantly helped us to understand the underlying pathological mechanisms of brain diseases [2,3,4,5], and, therefore, it also has been employed to diagnose ASD [6,7,8].
In neuroimaging, resting-state functional magnetic resonance imaging (RS-fMRI) utilizes blood oxygen level-dependent (BOLD) signals to explore the biomarkers of nervous system diseases [9,10,11,12]. Recently, the RS-fMRI-based ASD diagnosis has made significant progress [13,14,15,16]. For instance, Zhao et al. [14] presented a multi-view high-order functional connectivity network (FCN) based on the RS-fMRI data for ASD vs. the normal control (NC) classification. These approaches were designed based on single-site data, and, therefore, they cannot be generalized to other sites’ data because of the heterogeneity of the data from different sites. Additionally, the approaches developed from a small number of samples may present overfitting.
The RS-fMRI samples from multiple sites are somewhat heterogeneous due to the scanner type and imaging acquisition protocol. In particular, heterogeneity of the RS-fMRI samples tends to cause low generalization performance [17,18,19]. To deal with the problem caused by heterogeneity, many diagnostic modelings of ASD depending on samples from multiple sites have been explored [20,21]; these can be roughly categorized into two types. The first type ignores the heterogeneity [19,20] by assuming that samples from multiple sites are collected from the same or similar distribution. For instance, Brown et al. [21] proposed the element-wise layer for DNNs to predict ASD without considering the heterogeneity of data from different sites. The other type aims to avoid the adverse effect of data heterogeneity on the results [22,23,24]. For example, Niu et al. [22] proposed a multi-channel deep attention neural network to capture the correlations in multi-site data by integrating multi-layer neural networks, attention mechanisms, and feature fusion. However, these methods need extensive data that are hard to collect.
To improve generalization on the limited number of RS-fMRI samples, this study presents a similarity measure-based method for early ASD diagnosis. First, a Siamese network is devised to reduce the negative effect of heterogeneity on the performance of the model. Afterward, we design an independent objective function for each training site, and the total objective function is summed to train the parameters based on the back-propagation algorithm. Finally, a few samples, which the model has never seen before, are used to fine-tune the parameters with the aim of making the model adapt to new samples.

2. Problem Formulation

To improve the generalization and accuracy of ASD diagnoses, we follow the few-shot learning idea [25,26] to design the similarity measure-based approach in this study. Specifically, few-shot learning is formulated as learning a classifier to recognize the remaining samples given a few labeled samples for each class in the target site. If the classifier is directly trained utilizing traditional optimization algorithms, it is hard to obtain satisfactory performance due to the heterogeneity. Under the few-shot learning strategy, some imaging sites are used to train the parameters, while others are used to fine-tune the model parameters and evaluate the performance of the model. Accordingly, the whole dataset used in this study is partitioned into training sites, target sites, and a baseline site (set). In particular, the training sets are split further into meta-training sets and meta-test sets, and we aim to match the samples of the target site to the baseline site accurately.

3. Materials and Methods

This section illustrates the proposed method in detail, especially the Siamese network and the training strategy.

3.1. Materials

In this work, we sampled the studied RS-fMRI data from the Autism Brain Imaging Data Exchange (ABIDE), which aims to facilitate discovery and comparison in ASD research. We included data from the C-PAC preprocessing pipeline of ABIDE [27]. The data were excluded when the scanning data was missing from the original imaging data at some time points. After preprocessing, the RS-fMRI data from 12 imaging sites (402 ASD patients and 423 NCs) were included in the final analysis. The detailed demographic information, including age and sex, is summarized in Table 1. In this work, the AAL atlas [28] with 116 brain regions is used for brain parcellation.

3.2. Constructing Dynamic FCN

Functional connectivity exploits the temporal correlation of BOLD signals in different brain regions to demonstrate how structural segregation and functionally specialized brain regions interact. FCNs are of great significance for discovering the functional organization of the human brain and for finding biomarkers for neuropsychiatric diseases [10,29,30]. Studies have shown that dynamic functional connectivity patterns have significant contributions to the diagnosis of neurological diseases [9,31,32].
The sliding window strategy is a popular method for constructing a dynamic FCN (D-FCN). The illustration of constructing the D-FCN is shown in Figure 1. A D-FCN is a sequential collection of sub-networks created by dividing the entire RS-fMRI time series into multiple overlapping sub-segments, each of which is constructed as a sub-network that reflects short-term correlations. The short-term correlation in k-th window is calculated by Pearson’s correlation coefficient, as follows:
F C i j k = m = 1 M x i m ( k ) x ¯ i ( k ) x j m ( k ) x ¯ j ( k ) m = 1 M x i m ( k ) x ¯ i ( k ) 2 m = 1 M x j m ( k ) x ¯ j ( k ) 2
where M is the length of a segment, x i ( k ) denotes the BOLD signals of the i-th ROI, x ¯ i ( k ) is the average value of all elements in x i ( k ) , and x i m ( k ) is the m-th element in x i ( k ) . Thus, a sub-FCN of a D-FCN is constructed as D ( k ) = F C i j ( k ) 1 k K , and the corresponding D-FCN can be represented as D = [ D ( 1 ) , , D ( k ) , , D ( K ) ] , where K is the total number of segments.
In this work, the mean matrix of the D-FCN [14,15,33] is used as the input. Taking into account the symmetry of the functional connectivity matrix, the off-diagonal triangular parts on it are vectored as the feature vector.

3.3. Siamese Network Framework

Recently, deep neural networks have been used broadly in various fields [34,35,36,37,38]. However, training neural networks requires abundant data. For some fields, it is difficult to collect a large amount of data. Therefore, few-shot learning [25,26] is proposed to solve this problem, and reliable results have been obtained in multitudinous studies. Few-shot learning is usually implemented based on metric learning methods, such as Prototype networks and Siamese networks.
As shown in Figure 2, the Siamese network [39,40] consists of two sub-networks that share parameters. In this work, the Siamese network consists of two identical feature extractors, an autoencoder, and fully connected layers; the input is a 6670-dimensional feature vector extracted from each subject, and the output is the similarity of two subjects. The calculation similarity unit is formulated as:
d i s t x ^ 1 , x ^ 2 = | | x ^ 1 x ^ 2 | | 1
where x ^ 1 and x ^ 2 are the outputs of the two feature extractors, respectively. In order to intuitively compare, the sigmoid function is used to map the d i s t x ^ 1 , x ^ 2 within the range (0, 1).
The loss function is important for artificial neural networks to generate separable representations for unseen classes. In this work, the mean square error is employed as the loss function:
L = 1 2 d i s t ( x ^ 1 , x ^ 2 ) y 2
where y is the label, indicating whether the two subjects are from the same category, that is:
y = 0 , if the two subjects are from the same category 1 , otherwise
As illustrated in Figure 2, the training Siamese network needs to input paired samples and optimize the model parameters by minimizing the loss function. In the settings of this section, the similarity is measured by the distance (Equation (2)), so when minimizing the loss function (Equation (3)), the distance between paired samples from the same category should be small, and the distance between paired samples from different categories should be large.
In the test phase, the test dataset is divided into a support set and a query set. Suppose that there are C categories in the data set and N samples in the support set of each category. This few-shot learning classification task is called a C-way N-shot task. Figure 3 shows the test flowchart of the trained Siamese network model. In our case, we have two categories (NC and ASD), so it is a two-way N-shot task. Given a query sample Q from the query set, each support set has N support samples. The distances/similarities between the query sample Q and the samples in the two support sets are calculated to predict the label of sample Q.

3.4. Few-Shot Training Strategy

Because the underlying pathology of the ASD cross-sites is similar, it can be reasonably assumed that the data extracted from multiple sites share an inherent underlying data structure [23]. The baseline set is used in both the training and final performance test. In other words, we use the data of a specific site as the baseline and compare the subjects from other imaging sites with the baseline. Our goal is to train the model parameters so that the distance/similarity comparison between other site data and baseline site data can achieve satisfactory results. That is to say, in addition to training the model parameters, a site should be found as the baseline site during the training of the model, which will be used as the support set in the test phase.
Unlike the traditional few-shot task, called the C-way N-shot task, a specific site is selected as the baseline set. It is worth noting that the number of NC samples in the baseline set may not be equal to the number of ASD samples. Therefore, it can be called a C-way (C = 2 in our case) task, but the N is uncertain. In this work, we employ the prototype network [41] for reference to convert the few-shot task into a two-way one-shot task. Specifically, the average abstract features (extracted by feature extractor) of the NC/ASD samples in the baseline set as the prototypical features are used to calculate the similarity with other samples.
During training, the data for each site for training is partitioned into a meta-training set and meta-test set. The training process is shown in Figure 4. The model parameters are determined by minimizing the loss between the meta-training set and the baseline set; the meta-test set is used to verify the performance of the model in each iteration. Each training site will obtain an independent loss. Suppose that there are K sites for training, the sum of K losses is used as the total loss for the tuning parameters:
L total = k = 1 K L k
Using the total loss as the objective function can prevent the model from overfitting the data of a certain site during the training process and promote the model’s learning of the “basic concept” of the similarity.

3.5. Experimental Setup

We selected five imaging sites as the baseline set and training sets, UCLA, UM, USM, NYU, and YALE. In each training set, 70% of the samples are used to train the model parameters, and the remaining 30% are used to evaluate the performance of the model during training. One of the five sites is selected as the baseline set, and the other four sites are selected to be the training sets during the training process. Determine the model parameters by minimizing the loss between the baseline set and the training sets.
Additionally, another seven imaging sites are used to test the performance of the final model, which are independent of the training sets. The test strategy is to predict the category by comparing the similarity between the target subject and the NC and ASD samples in the baseline set. Specifically, if the similarity between the target subject and the NC samples in the baseline set is higher, the label is predicted to be NC; otherwise, the label is ASD.
In this article, the accuracy, precision, and F1 score are used as the criteria for experimental performance evaluation.

4. Results

4.1. Classification Performance on Meta-Test Sets

When one site was selected as the baseline site, the average meta-test performance of the other four sites is presented in Table 2, and, in turn, each site except for the target sites was selected as the baseline site.
From Table 2, we can make the following observations. First, the selected baseline site affects the performance of the model. Due to the heterogeneity of the data among different imaging sites, it is difficult for the learner to make the loss function of each site converge to a certain range, although the total loss may converge. Therefore, when different baseline sets are selected, the learner will update the model parameters through different loss values. As a result, changing the baseline set will change the classification performance of the model. Second, the proposed model is effective on the training sites. Although the meta-test sets are selected from the training sets, they are not used to update and tune the model parameters. Each accuracy is greater than 60% in terms of classification accuracy, indicating that the model is effective on training sites. Finally, from the results, the average accuracy is between 60% and 70%, which can reasonably be considered to mean that the model does not overfit or underfit the data in training sites.
The loss curves are shown in Figure 5. The training loss is based on the following experimental settings. The RS-fMRI data from the YALE imaging site are selected as the baseline set; the data from the UCLA, UM, USM, and NYU imaging sites are selected as training sets. As can be seen from Figure 5, the loss function converges well.

4.2. Generalization Performance on Target Sites

To validate the generalization performance of the proposed model, we compared the model with several classical methods including the support vector machine (SVM), stacked autoencoder (SAE), and random forest (RF). The SVM and RF are implemented using the Python-based sklearn library, and default parameters are used. The SAE consists of three full connection layers with tanh activation function and uses two full connection layers for classification.
In the current experiment, the RS-fMRI data from the YALE imaging site are selected as the baseline set for the proposed method, and several subjects’ data from each target site are selected to fine-tune the model parameters. In order to avoid the problem of a sample imbalance, some data are randomly removed from the categories with plenty of samples to make equal the number of ASD and NC samples. For consistency comparison, the RS-fMRI data from the five imaging sites (i.e., UCLA, UM, USM, NYU, and YALE) were combined for training the SVM, RF, and SAE. The data from the target site are used as a test set to evaluate performance. In the training process, 70% of the training samples are used to train the comparison methods’ parameters, and the remaining 30% are used to evaluate the performances of the models during training. The results are summarized in Table 3. From the results, the ability of the proposed method to generalize to other imaging sites outperforms the competing methods.

5. Discussion

In light of their simplicity and ease of implementation, classic machine learning methods, such as the SVM and RF, have been widely used in previous RS-fMRI studies. However, to avert inter-site heterogeneity, these methods are often modeled and validated based on single-site data, which restricts their generalization to other imaging sites. Since the SAE contributes to easy overfitting when dealing with small sample size data, we merge the data from the training sites for training the SAE. In general, the SAE reshapes feature patterns in vector form to learn more informative high-level features for diagnosing ASD, but it is still difficult to generalize to other imaging sites in the face of heterogeneous data. Few-shot learning occurs by transferring knowledge learned during training tasks to unseen tasks. In our case, we chose four training sites and one baseline site, and we expected the model to learn general concepts to distinguish ASD from NC. This general capability may not be sufficient to deal with some unique characteristics of the target site, as fine-tuning of model parameters is required using a small amount of data from the query site. From the experimental results summarized in Table 3, the generalization performance of the proposed model outperforms the comparison methods, indicating that our method is effective for classifying ASD based on multi-site data.
It can be seen from Figure 5 that, after the training step is about 100, the loss value tends to be stable. This shows that the proposed method can converge rapidly. In addition, the trends of the four curves are similar, and there is no huge difference in the loss value. Except for one curve, the other three curves are almost coincidental. This shows that the model is not overfitted to the particular training station and is balanced on multiple training sites.

6. Conclusions

This paper presents a similarity measure-based approach for ASD diagnosis with RS-fMRI data. A unique property of this study is that samples from multiple sites are used to co-learn a similarity function with the baseline site, enabling the presented approach to achieve good generalization on unseen samples from target sites. Extensive experiments on the ABIDE I dataset show that the proposed approach has a robust generalization performance with comparable diagnostic accuracy in comparison to several well-established methods.
In a previous study, the ComBat harmonization method was used to eliminate the impact of data heterogeneity among sites [42]. In future work, we consider combining small sample learning and ComBat to better eliminate the adverse impact of data heterogeneity between sites. In addition, this work has yielded satisfactory results only using RS-fMRI data, but some recent studies have shown that it is possible to use multimodal data to diagnose neurological diseases, such as combining RS-fMRI data and structural MRI data for disease diagnosis. Using multimodal data to diagnose neurological diseases will be the focus of our future work.

Author Contributions

Conceptualization, Q.Z.; Methodology, X.Z.; Software, H.Y.; Formal analysis, X.Z.; Investigation, X.Z.; Data curation, Z.W.; Writing—original draft, X.Z.; Writing—review & editing, S.P.S. and H.Y.; Project administration, Q.Z.; Funding acquisition, Z.W. and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the Key Research and Development Project of Hainan Province (Grant No. ZDYF2021GXJS017 and No. ZDYF2021GXJS016), the National Natural Science Foundation of China (Grant No. 6216070295 and No. 82160345), and the Key Science and Technology Plan Project of Haikou (Grant No. 2021-016).

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to the data used in this study is from an public database.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

In this study, the data is from a public database, which is Autism Brain Imaging Data Exchange (ABIDE). The data is available at http://fcon_1000.projects.nitrc.org/indi/abide/ accessed on 24 October 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ecker, C.; Marquand, A.; Mourão-Miranda, J.; Johnston, P.; Daly, E.M.; Brammer, M.J.; Maltezos, S.; Murphy, C.M.; Robertson, D.; Williams, S.C.; et al. Describing the brain in autism in five dimensions—Magnetic resonance imaging-assisted diagnosis of autism spectrum disorder using a multiparameter classification approach. J. Neurosci. 2010, 30, 10612–10623. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Li, W.; Qiao, L.; Zhang, L.; Wang, Z.; Shen, D. Functional Brain Network Estimation With Time Series Self-Scrubbing. IEEE J. Biomed. Health Inform. 2019, 23, 2494–2504. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, Z.; Zhu, X.; Adeli, E.; Zhu, Y.; Nie, F.; Munsell, B.; Wu, G. Multi-modal classification of neurodegenerative disease by progressive graph-based transductive learning. Med. Image Anal. 2017, 39, 218–230. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhao, F.; Han, Z.; Cheng, D.; Mao, N.; Chen, X.; Li, Y.; Fan, D.; Liu, P. Hierarchical Synchronization Estimation of Low-and High-Order Functional Connectivity Based on Sub-Network Division for the Diagnosis of Autism Spectrum Disorder. Front. Neurosci. 2021, 15, 810431. [Google Scholar] [CrossRef]
  5. Zhang, X.; Shams, S.P.; Yu, H.; Wang, Z.; Zhang, Q. A pairwise functional connectivity similarity measure method based on few-shot learning for early MCI detection. Front. Neurosci. 2022, 16, 1081788. [Google Scholar] [CrossRef]
  6. Li, W.; Wang, Z.; Zhang, L.; Qiao, L.; Shen, D. Remodeling pearson’s correlation for functional brain network estimation and autism spectrum disorder identification. Front. Neuroinform. 2017, 11, 55. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, J.; Wang, Q.; Zhang, H.; Chen, J.; Wang, S.; Shen, D. Sparse Multiview Task-Centralized Ensemble Learning for ASD Diagnosis Based on Age- and Sex-Related Functional Connectivity Patterns. IEEE Trans. Cybern. 2019, 49, 3141–3154. [Google Scholar] [CrossRef]
  8. Zhao, F.; Li, N.; Pan, H.; Chen, X.; Li, Y.; Zhang, H.; Mao, N.; Cheng, D. Multi-View Feature Enhancement Based on Self-Attention Mechanism Graph Convolutional Network for Autism Spectrum Disorder Diagnosis. Front. Hum. Neurosci. 2022, 16, 918969. [Google Scholar] [CrossRef]
  9. Zhao, F.; Chen, Z.; Rekik, I.; Lee, S.W.; Shen, D. Diagnosis of autism spectrum disorder using central-moment features from low-and high-order dynamic resting-state functional connectivity networks. Front. Neurosci. 2020, 14, 258. [Google Scholar] [CrossRef]
  10. Zhao, F.; Zhang, H.; Rekik, I.; An, Z.; Shen, D. Diagnosis of autism spectrum disorders using multi-level high-order functional networks derived from resting-state functional mri. Front. Hum. Neurosci. 2018, 12, 184. [Google Scholar] [CrossRef] [Green Version]
  11. Zhang, G.; Cai, B.; Zhang, A.; Stephen, J.M.; Wilson, T.W.; Calhoun, V.D.; Wang, Y.P. Estimating Dynamic Functional Brain Connectivity With a Sparse Hidden Markov Model. IEEE Trans. Med. Imaging 2020, 39, 488–498. [Google Scholar] [CrossRef] [PubMed]
  12. Ali, R.; Li, H.; Dillman, J.R.; Altaye, M.; Wang, H.; Parikh, N.A.; He, L. A self-training deep neural network for early prediction of cognitive deficits in very preterm infants using brain functional connectome data. Pediatr. Radiol. 2022, 52, 2227–2240. [Google Scholar] [CrossRef] [PubMed]
  13. Zhao, F.; Chen, Z.; Rekik, I.; Liu, P.; Mao, N.; Lee, S.W.; Shen, D. A novel unit-based personalized fingerprint feature selection strategy for dynamic functional connectivity networks. Front. Neurosci. 2021, 15, 651574. [Google Scholar] [CrossRef] [PubMed]
  14. Zhao, F.; Zhang, X.; Thung, K.H.; Mao, N.; Lee, S.W.; Shen, D. Constructing Multi-View High-Order Functional Connectivity Networks for Diagnosis of Autism Spectrum Disorder. IEEE Trans. Biomed. Eng. 2022, 69, 1237–1250. [Google Scholar] [CrossRef]
  15. Xie, Q.; Zhang, X.; Rekik, I.; Chen, X.; Mao, N.; Shen, D.; Zhao, F. Constructing high-order functional connectivity network based on central moment features for diagnosis of autism spectrum disorder. PeerJ 2021, 9, e11692. [Google Scholar] [CrossRef]
  16. Kong, Y.; Gao, J.; Xu, Y.; Pan, Y.; Wang, J.; Liu, J. Classification of autism spectrum disorder by combining brain connectivity and deep neural network classifier. Neurocomputing 2019, 324, 63–68. [Google Scholar] [CrossRef]
  17. Liu, X.; Huang, H. Alterations of functional connectivities associated with autism spectrum disorder symptom severity: A multi-site study using multivariate pattern analysis. Sci. Rep. 2020, 10, 4330. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, F.; Qiao, L.; Shi, F.; Yap, P.T.; Shen, D. Feature fusion via hierarchical supervised local CCA for diagnosis of autism spectrum disorder. Brain Imaging Behav. 2017, 11, 1050–1060. [Google Scholar] [CrossRef]
  19. Huang, H.; Liu, X.; Jin, Y.; Lee, S.W.; Wee, C.Y.; Shen, D. Enhancing the representation of functional connectivity networks by fusing multi-view information for autism spectrum disorder diagnosis. Hum. Brain Mapp. 2019, 40, 833–854. [Google Scholar] [CrossRef] [Green Version]
  20. Heinsfeld, A.S.; Franco, A.R.; Craddock, R.C.; Buchweitz, A.; Meneguzzi, F. Identification of autism spectrum disorder using deep learning and the ABIDE dataset. NeuroImage Clin. 2018, 17, 16–23. [Google Scholar] [CrossRef]
  21. Brown, C.J.; Kawahara, J.; Hamarneh, G. Connectome priors in deep neural networks to predict autism. In Proceedings of the 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 110–113. [Google Scholar]
  22. Niu, K.; Guo, J.; Pan, Y.; Gao, X.; Peng, X.; Li, N.; Li, H. Multichannel deep attention neural networks for the classification of autism spectrum disorder using neuroimaging and personal characteristic data. Complexity 2020, 2020, 1357853. [Google Scholar] [CrossRef]
  23. Wang, M.; Zhang, D.; Huang, J.; Yap, P.T.; Shen, D.; Liu, M. Identifying Autism Spectrum Disorder With Multi-Site fMRI via Low-Rank Domain Adaptation. IEEE Trans. Med. Imaging 2020, 39, 644–655. [Google Scholar] [CrossRef] [PubMed]
  24. Eslami, T.; Mirjalili, V.; Fong, A.; Laird, A.R.; Saeed, F. ASD-DiagNet: A hybrid learning approach for detection of autism spectrum disorder using fMRI data. Front. Neuroinformatics 2019, 13, 70. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Ma, Y.; Zhao, S.; Wang, W.; Li, Y.; King, I. Multimodality in meta-learning: A comprehensive survey. Knowl.-Based Syst. 2022, 250, 108976. [Google Scholar] [CrossRef]
  26. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. 2020, 53, 63. [Google Scholar] [CrossRef]
  27. Craddock, C.; Benhajali, Y.; Chu, C.; Chouinard, F.; Evans, A.; Jakab, A.; Khundrakpam, B.S.; Lewis, J.D.; Li, Q.; Milham, M.; et al. The neuro bureau preprocessing initiative: Open sharing of preprocessed neuroimaging data and derivatives. Front. Neuroinform. 2013, 7, 27. [Google Scholar]
  28. Tzourio-Mazoyer, N.; Landeau, B.; Papathanassiou, D.; Crivello, F.; Etard, O.; Delcroix, N.; Mazoyer, B.; Joliot, M. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage 2002, 15, 273–289. [Google Scholar] [CrossRef] [PubMed]
  29. Shuai, X.X.; Kong, X.C.; Zou, Y.; Wang, S.Q.; Wang, Y.H. Global Functional Network Connectivity Disturbances in Parkinson’s Disease with Mild Cognitive Impairment by Resting-State Functional MRI. Curr. Med. Sci. 2020, 40, 1057–1066. [Google Scholar] [CrossRef]
  30. Dai, L.; Yu, Y.; Zhao, H.; Zhang, X.; Su, Y.; Wang, X.; Hu, S.; Dai, H.; Hu, C.; Ke, J. Altered local and distant functional connectivity density in chronic migraine: A resting-state functional MRI study. Neuroradiology 2021, 63, 555–562. [Google Scholar] [CrossRef]
  31. Zhu, Y.; Li, X.; Sun, Y.; Wang, H.; Guo, H.; Sui, J. Investigating Neural Substrates of Individual Independence and Interdependence Orientations via Efficiency-Based Dynamic Functional Connectivity: A Machine Learning Approach. IEEE Trans. Cogn. Dev. Syst. 2022, 14, 761–771. [Google Scholar] [CrossRef]
  32. Dai, M.; Zhang, Z.; Srivastava, A. Analyzing Dynamical Brain Functional Connectivity as Trajectories on Space of Covariance Matrices. IEEE Trans. Med. Imaging 2020, 39, 611–620. [Google Scholar] [CrossRef] [PubMed]
  33. Zhou, Y.; Qiao, L.; Li, W.; Zhang, L.; Shen, D. Simultaneous estimation of low-and high-order functional connectivity for identifying mild cognitive impairment. Front. Neuroinformatics 2018, 12, 3. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Hu, X.; Ding, X.; Bai, D.; Zhang, Q. A Compressed Model-Agnostic Meta-Learning Model Based on Pruning for Disease Diagnosis. J. Circuits Syst. Comput. 2022, 32, 2350022. [Google Scholar] [CrossRef]
  35. Zhu, R.; Jiang, Q.; Huang, X.; Li, D.; Yang, Q. A Reinforcement-Learning-Based Opportunistic Routing Protocol for Energy-Efficient and Void-Avoided UASNs. IEEE Sens. J. 2022, 22, 13589–13601. [Google Scholar] [CrossRef]
  36. Yu, H.; Yang, L.T.; Zhang, Q.; Armstrong, D.; Deen, M.J. Convolutional neural networks for medical image analysis: State-of-the-art, comparisons, improvement and perspectives. Neurocomputing 2021, 444, 92–110. [Google Scholar] [CrossRef]
  37. Yu, H.; Yang, L.T.; Fan, X.; Zhang, Q. A deep residual computation model for heterogeneous data learning in smart Internet of Things. Appl. Soft Comput. 2021, 107, 107361. [Google Scholar] [CrossRef]
  38. Wang, S.; Wang, S.; Liu, Z.; Zhang, Q. A role distinguishing Bert model for medical dialogue system in sustainable smart city. Sustain. Energy Technol. Assess. 2023, 55, 102896. [Google Scholar] [CrossRef]
  39. Bromley, J.; Guyon, I.; LeCun, Y.; Säckinger, E.; Shah, R. Signature verification using a “siamese” time delay neural network. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA; 1993; Volume 6, pp. 669–688. [Google Scholar]
  40. Chicco, D. Siamese neural networks: An overview. In Artificial Neural Networks; Humana: New York, NY, USA, 2021; pp. 73–94. [Google Scholar]
  41. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  42. Yu, M.; Linn, K.A.; Cook, P.A.; Phillips, M.L.; McInnis, M.; Fava, M.; Trivedi, M.H.; Weissman, M.M.; Shinohara, R.T.; Sheline, Y.I. Statistical harmonization corrects site effects in functional connectivity measurements from multi-site fMRI data. Hum. Brain Mapp. 2018, 39, 4213–4227. [Google Scholar] [CrossRef]
Figure 1. The illustration of constructing the D-FCN. The F C i j ( k ) 1 k K denotes the temporal correlation of the i-th and j-th ROI in the k-th segment.
Figure 1. The illustration of constructing the D-FCN. The F C i j ( k ) 1 k K denotes the temporal correlation of the i-th and j-th ROI in the k-th segment.
Diagnostics 13 00218 g001
Figure 2. Illustration of the Siamese network.
Figure 2. Illustration of the Siamese network.
Diagnostics 13 00218 g002
Figure 3. Test flowchart of the Siamese network.
Figure 3. Test flowchart of the Siamese network.
Diagnostics 13 00218 g003
Figure 4. Training flowchart of the Siamese network. The prototypical features are extracted from the baseline site.
Figure 4. Training flowchart of the Siamese network. The prototypical features are extracted from the baseline site.
Diagnostics 13 00218 g004
Figure 5. The training losses. The l o s s 0 , l o s s 1 , l o s s 2 , and l o s s 3 are the training losses on the UCLA, UM, USM, and NYU sites, respectively.
Figure 5. The training losses. The l o s s 0 , l o s s 1 , l o s s 2 , and l o s s 3 are the training losses on the UCLA, UM, USM, and NYU sites, respectively.
Diagnostics 13 00218 g005
Table 1. Demographic information of the subjects from 12 imaging sites. Average (SD) (SD, standard deviation); M/F: male/female.
Table 1. Demographic information of the subjects from 12 imaging sites. Average (SD) (SD, standard deviation); M/F: male/female.
 ASDNC
SiteAge Avg (SD)Count: M/FAge Avg (SD)Count: M/F
KKI10.2 (1.5)12 / 310.1 (1.2)18 / 7
OHSU11.4 (2.2)12 / 010.1 (1.1)14 / 0
OLIN16.5 (3.4)16 / 316.7 (3.6)13 / 2
TRINITY16.8 (3.2)22 / 017.0 (3.8)24 / 0
MAXMUN26.1 (14.9)21 / 324.5 (9.1)25 / 1
LEUVEN17.8 (5.0)26 / 318.3 (5.1)28 / 5
PITT19.8 (8.5)16 / 318.0 (6.7)14 / 5
UCLA13.0 (2.5)48 / 613.0 (1.9)38 / 6
UM13.0 (2.3)55 / 914.9 (3.6)56 / 17
USM23.0 (7.4)44 / 021.3 (8.4)25 / 0
NYU14.4 (6.9)62 / 1015.6 (6.1)72 / 25
YALE12.7 (3.0)20 / 812.7 (2.8)20 / 8
Table 2. The performance on meta-test sets.
Table 2. The performance on meta-test sets.
Baseline SiteAccuracy (%)Precision (%)F1 Score (%)
UCLA61.30 ± 1.6160.87 ± 1.6062.00 ± 3.76
UM68.83 ± 4.7567.35 ± 5.4970.51 ± 3.27
USM65.82 ± 3.8466.41 ± 4.4365.25 ± 4.08
NYU65.19 ± 6.1263.27 ± 6.7369.00 ± 4.67
YALE69.59 ± 1.4472.16 ± 7.0068.42 ± 3.51
Table 3. Performance comparisons between the proposed model and competing methods.
Table 3. Performance comparisons between the proposed model and competing methods.
SiteMethodAccuracy (%)Precision (%)F1 Score (%)
TRINITYRF61.3664.7156.41
SVM68.1870.0066.67
SAE68.1872.2265.00
Ours73.6884.2180.00
OLINRF73.3373.3373.33
SVM73.3370.5975.00
SAE73.3373.3373.33
Ours80.7676.9278.57
OHSURF33.3333.3333.33
SVM54.1753.3359.26
SAE41.6740.0036.36
Ours57.1457.1457.14
PITTRF52.6354.5540.00
SVM50.0050.0034.48
SAE55.2657.1448.48
Ours57.6992.3075.00
LEUVENRF53.4562.5027.03
SVM58.6285.7133.33
SAE65.5284.6252.38
Ours71.1588.4682.35
MAXMUNRF47.9147.6244.44
SVM56.2557.8951.16
SAE60.4264.7153.66
Ours61.9076.1966.66
KKIRF60.0061.5357.14
SVM66.6664.7168.75
SAE63.3364.2962.07
Ours66.6675.0070.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Shams, S.P.; Yu, H.; Wang, Z.; Zhang, Q. A Similarity Measure-Based Approach Using RS-fMRI Data for Autism Spectrum Disorder Diagnosis. Diagnostics 2023, 13, 218. https://doi.org/10.3390/diagnostics13020218

AMA Style

Zhang X, Shams SP, Yu H, Wang Z, Zhang Q. A Similarity Measure-Based Approach Using RS-fMRI Data for Autism Spectrum Disorder Diagnosis. Diagnostics. 2023; 13(2):218. https://doi.org/10.3390/diagnostics13020218

Chicago/Turabian Style

Zhang, Xiangfei, Shayel Parvez Shams, Hang Yu, Zhengxia Wang, and Qingchen Zhang. 2023. "A Similarity Measure-Based Approach Using RS-fMRI Data for Autism Spectrum Disorder Diagnosis" Diagnostics 13, no. 2: 218. https://doi.org/10.3390/diagnostics13020218

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop