Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Dec 1:283:120412.
doi: 10.1016/j.neuroimage.2023.120412. Epub 2023 Oct 18.

Neuroimaging-based classification of PTSD using data-driven computational approaches: A multisite big data study from the ENIGMA-PGC PTSD consortium

Xi Zhu  1 Yoojean Kim  2 Orren Ravid  2 Xiaofu He  3 Benjamin Suarez-Jimenez  4 Sigal Zilcha-Mano  5 Amit Lazarov  6 Seonjoo Lee  1 Chadi G Abdallah  7 Michael Angstadt  8 Christopher L Averill  7 C Lexi Baird  9 Lee A Baugh  10 Jennifer U Blackford  11 Jessica Bomyea  12 Steven E Bruce  13 Richard A Bryant  14 Zhihong Cao  15 Kyle Choi  12 Josh Cisler  16 Andrew S Cotton  17 Judith K Daniels  18 Nicholas D Davenport  19 Richard J Davidson  20 Michael D DeBellis  9 Emily L Dennis  21 Maria Densmore  22 Terri deRoon-Cassini  23 Seth G Disner  19 Wissam El Hage  24 Amit Etkin  25 Negar Fani  26 Kelene A Fercho  27 Jacklynn Fitzgerald  28 Gina L Forster  29 Jessie L Frijling  30 Elbert Geuze  31 Atilla Gonenc  32 Evan M Gordon  33 Staci Gruber  32 Daniel W Grupe  20 Jeffrey P Guenette  34 Courtney C Haswell  9 Ryan J Herringa  35 Julia Herzog  36 David Bernd Hofmann  37 Bobak Hosseini  38 Anna R Hudson  39 Ashley A Huggins  9 Jonathan C Ipser  40 Neda Jahanshad  41 Meilin Jia-Richards  42 Tanja Jovanovic  43 Milissa L Kaufman  44 Mitzy Kennis  31 Anthony King  8 Philipp Kinzel  45 Saskia B J Koch  46 Inga K Koerte  45 Sheri M Koopowitz  40 Mayuresh S Korgaonkar  47 John H Krystal  48 Ruth Lanius  49 Christine L Larson  50 Lauren A M Lebois  51 Gen Li  52 Israel Liberzon  53 Guang Ming Lu  54 Yifeng Luo  15 Vincent A Magnotta  55 Antje Manthey  56 Adi Maron-Katz  25 Geoffery May  57 Katie McLaughlin  58 Sven C Mueller  39 Laura Nawijn  59 Steven M Nelson  60 Richard W J Neufeld  22 Jack B Nitschke  20 Erin M O'Leary  17 Bunmi O Olatunji  61 Miranda Olff  30 Matthew Peverill  62 K Luan Phan  63 Rongfeng Qi  54 Yann Quidé  64 Ivan Rektor  65 Kerry Ressler  51 Pavel Riha  65 Marisa Ross  66 Isabelle M Rosso  51 Lauren E Salminen  41 Kelly Sambrook  62 Christian Schmahl  36 Martha E Shenton  67 Margaret Sheridan  68 Chiahao Shih  17 Maurizio Sicorello  36 Anika Sierk  56 Alan N Simmons  69 Raluca M Simons  70 Jeffrey S Simons  70 Scott R Sponheim  71 Murray B Stein  12 Dan J Stein  40 Jennifer S Stevens  26 Thomas Straube  37 Delin Sun  9 Jean Théberge  22 Paul M Thompson  41 Sophia I Thomopoulos  41 Nic J A van der Wee  72 Steven J A van der Werff  72 Theo G M van Erp  73 Sanne J H van Rooij  26 Mirjam van Zuiden  30 Tim Varkevisser  31 Dick J Veltman  59 Robert R J M Vermeiren  72 Henrik Walter  56 Li Wang  74 Xin Wang  17 Carissa Weis  23 Sherry Winternitz  44 Hong Xie  17 Ye Zhu  52 Melanie Wall  1 Yuval Neria  3 Rajendra A Morey  9
Affiliations

Neuroimaging-based classification of PTSD using data-driven computational approaches: A multisite big data study from the ENIGMA-PGC PTSD consortium

Xi Zhu et al. Neuroimage. .

Abstract

Background: Recent advances in data-driven computational approaches have been helpful in devising tools to objectively diagnose psychiatric disorders. However, current machine learning studies limited to small homogeneous samples, different methodologies, and different imaging collection protocols, limit the ability to directly compare and generalize their results. Here we aimed to classify individuals with PTSD versus controls and assess the generalizability using a large heterogeneous brain datasets from the ENIGMA-PGC PTSD Working group.

Methods: We analyzed brain MRI data from 3,477 structural-MRI; 2,495 resting state-fMRI; and 1,952 diffusion-MRI. First, we identified the brain features that best distinguish individuals with PTSD from controls using traditional machine learning methods. Second, we assessed the utility of the denoising variational autoencoder (DVAE) and evaluated its classification performance. Third, we assessed the generalizability and reproducibility of both models using leave-one-site-out cross-validation procedure for each modality.

Results: We found lower performance in classifying PTSD vs. controls with data from over 20 sites (60 % test AUC for s-MRI, 59 % for rs-fMRI and 56 % for d-MRI), as compared to other studies run on single-site data. The performance increased when classifying PTSD from HC without trauma history in each modality (75 % AUC). The classification performance remained intact when applying the DVAE framework, which reduced the number of features. Finally, we found that the DVAE framework achieved better generalization to unseen datasets compared with the traditional machine learning frameworks, albeit performance was slightly above chance.

Conclusion: These results have the potential to provide a baseline classification performance for PTSD when using large scale neuroimaging datasets. Our findings show that the control group used can heavily affect classification performance. The DVAE framework provided better generalizability for the multi-site data. This may be more significant in clinical practice since the neuroimaging-based diagnostic DVAE classification models are much less site-specific, rendering them more generalizable.

Keywords: Classification; Deep learning; Machine learning; Multimodal MRI; Posttraumatic stress disorder.

PubMed Disclaimer

Conflict of interest statement

Declaration of Competing Interest Dr. Thompson received partial grant support from Biogen, Inc., and Amazon, Inc., for work unrelated to the current study; Dr. Lebois reports unpaid membership on the Scientific Committee for International Society for the Study of Trauma and Dissociation (ISSTD), grant support from the National Institute of Mental Health, K01 MH118467 and the Julia Kasparian Fund for Neuroscience Research, McLean Hospital. Dr. Lebois also reports spousal IP payments from Vanderbilt University for technology licensed to Acadia Pharmaceuticals unrelated to the present work. ISSTD and NIMH were not involved in the analysis or preparation of the manuscript; Dr. Etkin reports salary and equity from Alto Neuroscience, equity from Mindstrong Health and Akili Interactive. Other authors have no conflicts of interest to declare.

Figures

Fig. 1.
Fig. 1.
Brain features from structural MRI (s-MRI), resting state fMRI (rs-fMRI), and DTI (d-MRI) used in this study. T1-weighted images were processed using the FreeSurfer pipeline, the final s-MRI features included 96 ROI cortical thicknesses (CT) and volumes for both left and right hemispheres. Rs-fMRI images were preprocessed using ENIGMA HALFpipe workflow, the final rs-fMRI features included 10,878 ROI-to-ROI functional connectivity measures. DTI data were preprocessed following ENIGMA-DTI protocols, 42 Tract-Based Spatial Statistics (TBSS) derived features of mean FA were included in the analysis.
Fig. 2.
Fig. 2.
Overall analysis procedure. The analysis followed the structure of the three main aims and 2 supplementary aims of the paper. Goals 1 and 2 used data that was pooled across sites/scanners whereas goal 3 used site information to facilitate performance assessment. Goal 1, investigated both Support vector machine (SVM) and random forest (RF) for classification of s-MRI data, which was repeated for rs-fMRI, and then for D-MRI. Goal 2 investigated DVAE in conjunction with SVM and in conjunction with RF (DVAE+SVM/RF), first using s-MRI features and then using rs-fMRI features. Goal 3 investigated performance of SVM or RF on single site data that was tested separately on s-MRI, rs-fMRI, and D-MRI brain features. Goal 3 also investigated performance of LOSOCV tested separately on s-MRI (SVM, DVAE+SVM), rs-fMRI (SVM, DVAE+SVM), and D-MRI (SVM) brain features. S-MRI: structural MRI; RS-fMRI: resting state fMRI; D-MRI: diffusion MRI; SVM: support vector machine; RF: random forest; DVAE: Denoising variational autoencoder.
Fig. 3.
Fig. 3.
Denoising Variational Autoencoder analysis pipeline: The model was trained using rs-fMRI or s-MRI data from controls only. The samples were then split into a training+validation (70 %) and independent-test (30 %) data. Then 20 % of the training data was set aside for validation and hyperparameter tuning. Once the training+validation was completed, the model’s performance was evaluated on the independent-test data, which provides an unbiased estimate of how the model generalizes to unseen data. The resulting VAE model learned to encode healthy patterns from the input brain features into its latent representation. Later, the brain features from patients with PTSD (PTSD test set) were input into the same VAE model, and the latent variables were extracted as new features for classification analysis.
Fig. 4.
Fig. 4.
The overall classification performance (measured by cross validation AUC [CV AUC], and test AUC) between PTSD and all controls, between PTSD and HC, and between PTSD and TEHC, for s-MRI, rs-fMRI, and D-MRI. Error bar represents standard deviation of the 10 fold cross validation results.
Fig. 5.
Fig. 5.
Compare classification performance between PTSD and Controls using all features (labeled as RF or SVM) and DVAE-based latent variables (labeled as DVAE+RF or DVAE+SVM) in s-MRI (A) and rs-fMRI (B). Compared with the performance (CV AUC) using SVM of all features, the performance of DVAE+SVM (CV AUC) significantly improved for both s-MRI and rs-fMRI.
Fig. 6.
Fig. 6.
The reconstruction loss function of the Denoising Variational Autoencoder model for s-MRI (A), rs-fMRI (B), and D-MRI (C), blue line: loss for the training set (from control data), orange line: loss for the validation set (from control data), green line: loss for the validation data (from PTSD data).
Fig. 7.
Fig. 7.
s-MRI, rs-fMRI, and D-MRI single site performance for classification of PTSD from controls using SVM. The classification performance was measured by cross validation (CV) AUC, the dot indicates the average of the AUC of each fold in cross validation for each site, the line indicates the standard deviation of each fold in cross validation for each site. The boxplots were made by utilizing the boxplot() function from the seaborn library in Python. The box encompasses the interquartile interval, or the middle 50 % of the dataset. The upper and lower whiskers represent data points located in the top and bottom 25 % of the dataset. Data that fall outside this range are considered outliers and are plotted individually.
Fig. 8.
Fig. 8.
The comparison of Leave One Site Out Cross Validation (LOSOCV) performance with aggregated pooling method on the independent-test data using SVM for classification between PTSD and Controls across s-MRI (T1), rs-fMRI (RS), LOSOCV: In each iteration, one independent test site was completely left out from the training partition. Then the training set was further randomly partitioned into 10 subfolds for cross validation. Predictive performance was evaluated on the data from the hold-out site (presented light gray in the figure). Aggregated pooling method: data from all sites was included in the training process. We randomly split all imaging data into two subsets: 70 % of the data was used for training+validation (cross validation), and the remaining 30 % was used as independent test data. Random under-sampling was applied to the imbalanced groups, with the under-sampling transform applied to the training dataset on each split of a repeated 10-fold cross validation. Predictive performance was evaluated on the data from the independent test set (presented dark gray in the figure).

References

    1. Altmann A, Tolosi L, Sander O, et al., 2010. Permutation importance: a corrected feature importance measure. Bioinformatics 26, 1340–1347. - PubMed
    1. Arbabshirani MR, Plis S, Sui J, et al., 2017. Single subject prediction of brain disorders in neuroimaging: promises and pitfalls. Neuroimage 145, 137–165. - PMC - PubMed
    1. Bai X, Wang X, Liu X, et al., 2021. Explainable deep learning for efficient and robust pattern recognition: a survey of recent developments. Pattern Recognit. 120, 108102.
    1. Behzadi Y, Restom K, Liau J, et al., 2007. A component based noise correction method (CompCor) for BOLD and perfusion based fMRI. Neuroimage 37, 90–101. - PMC - PubMed
    1. Belleau EL, Ehret LE, Hanson JL, et al., 2020. Amygdala functional connectivity in the acute aftermath of trauma prospectively predicts severity of posttraumatic stress symptoms. Neurobiol. Stress 12, 100217. - PMC - PubMed

Publication types