Skip to main content
Full access
Articles
Published Date: 7 February 2024

Treatment Response Prediction in Major Depressive Disorder Using Multimodal MRI and Clinical Data: Secondary Analysis of a Randomized Clinical Trial

This article has been corrected.
VIEW CORRECTION
Publication: American Journal of Psychiatry

Abstract

Objective:

Response to antidepressant treatment in major depressive disorder varies substantially between individuals, which lengthens the process of finding effective treatment. The authors sought to determine whether a multimodal machine learning approach could predict early sertraline response in patients with major depressive disorder. They assessed the predictive contribution of MR neuroimaging and clinical assessments at baseline and after 1 week of treatment.

Methods:

This was a preregistered secondary analysis of data from the Establishing Moderators and Biosignatures of Antidepressant Response in Clinical Care (EMBARC) study, a multisite double-blind, placebo-controlled randomized clinical trial that included 296 adult outpatients with unmedicated recurrent or chronic major depressive disorder. MR neuroimaging and clinical data were collected before and after 1 week of treatment. Performance in predicting response and remission, collected after 8 weeks, was quantified using balanced accuracy (bAcc) and area under the receiver operating characteristic curve (AUROC) scores.

Results:

A total of 229 patients were included in the analyses (mean age, 38 years [SD=13]; 66% female). Internal cross-validation performance in predicting response to sertraline (bAcc=68% [SD=10], AUROC=0.73 [SD=0.03]) was significantly better than chance. External cross-validation on data from placebo nonresponders (bAcc=62%, AUROC=0.66) and placebo nonresponders who were switched to sertraline (bAcc=65%, AUROC=0.68) resulted in differences that suggest specificity for sertraline treatment compared with placebo treatment. Finally, multimodal models outperformed unimodal models.

Conclusions:

The study results confirm that early sertraline treatment response can be predicted; that the models are sertraline specific compared with placebo; that prediction benefits from integrating multimodal MRI data with clinical data; and that perfusion imaging contributes most to these predictions. Using this approach, a lean and effective protocol could individualize sertraline treatment planning to improve psychiatric care.
Major depressive disorder (MDD) is the second largest contributor to disability worldwide (1). To date, no objective predictor of individual treatment response exists (2), and first-step antidepressant therapy yields remission in only one-third of cases (3). Consequently, patients often undergo multiple sequential treatments and treatment combinations (4). Each unsuccessful treatment prolongs the disease burden and is associated with the risk of substantial side effects and high societal costs (5). To expedite remission, clinically valuable biomarkers are needed to indicate individual treatment response before or early after treatment initiation (6).
Neuroimaging using MRI has shown promising results in predicting treatment response in MDD (7). Although several predictive neuroimaging biomarkers have been proposed over the years, their clinical application is still far from standard practice because studies are highly heterogeneous, have small effect sizes, and are mainly based on unimodal neuroimaging measures (8). Therefore, it was recently suggested that individual neuroimaging predictors should be combined with other (e.g., clinical) predictors using machine learning to create a larger effect and increase the robustness of neuroimaging findings (9). However, only two of 27 studies combined multiple imaging modalities or clinical information to predict treatment response (10). A recent study by Sajjadian et al. (11) demonstrated the benefits of combining clinical and imaging data before and after 2 weeks of treatment in predicting escitalopram treatment response. Here, we take those findings a step further by incorporating all clinically applicable MRI modalities that have been shown to have predictive value (8) to develop a multimodal model based on a large randomized clinical trial with a placebo treatment arm that allows not only for external validation but also for testing for model specificity to sertraline treatment compared with placebo treatment.
The primary purpose of this study was to determine whether the response to sertraline treatment can be predicted before (pretreatment) or 1 week after treatment initiation (early treatment) using machine-learning-based integration of multiple MRI modalities and clinical assessment data. Our primary hypothesis was that prediction of sertraline treatment using a multimodal approach is significantly better than the best alternative, the a priori response rates (chance), and thus standard treatment planning. Furthermore, we investigated the specificity of our predictive model for sertraline compared with placebo treatment. We hypothesized that our prediction model would be sertraline specific, and thus we expected reduced test performance in placebo-treated patients but similar performance in sertraline-treated placebo nonresponders. Finally, we tested our hypothesis that a multimodal approach outperforms unimodal approaches.

Methods

Study Design

The study methods were preregistered prior to analysis at the Open Science Framework register (12). We performed a secondary analysis of data from the Establishing Moderators and Biosignatures of Antidepressant Response for Clinical Care for Depression (EMBARC) study, a multisite double-blind randomized clinical trial (ClinicalTrials.gov identifier, NCT01407094) funded by the National Institute of Mental Health. EMBARC was designed to identify pretreatment and early-treatment (week 1) clinical, neuroimaging, neurophysiological, and behavioral moderators for predicting response (at week 8) to sertraline and placebo in patients with MDD. Patients provided written informed consent after receiving a complete description of the study. Details of the trial design have been reported elsewhere (13).
The primary outcomes of EMBARC were based on symptom severity as measured with the 17-item Hamilton Depression Rating Scale (HAM-D). The trial consisted of two 8-week phases. In phase 1, all patients were randomized to treatment with either sertraline or placebo (see the supplemental methods section of the online supplement for the dosage regimen). In phase 2, nonresponders to placebo were switched to sertraline.

Population

EMBARC included 296 outpatients 18–65 years of age who were diagnosed as having recurrent or chronic MDD and were not taking medication for depression; participants were not required to be medication naive. To test our hypotheses, we defined three population subgroups, for each of which we trained four sets of models. A flow chart of these subgroups is provided in Figure 1. Subgroup A consisted of all 146 patients randomized to the sertraline treatment arm, and subgroup B the 150 patients randomized to the placebo treatment arm in phase 1 of the study. Subgroup C comprised 77 patients from subgroup B who did not respond to placebo treatment in phase 1 of the study, after which they were switched to sertraline in phase 2.
FIGURE 1. CONSORT flow diagram for a study of treatment response prediction in major depressive disorder using multimodal MRI and clinical dataa
aThe EMBARC clinical trial consists of two study arms and two 8-week phases. One arm was randomized to sertraline treatment and the other to placebo. Highlighted are three population subgroups used in our analyses. The first two subgroups, in phase 1 of the study, are those treated with sertraline (A) or placebo (B). The third subgroup consists of patients who did not respond to placebo treatment in phase 1 and were switched to treatment with sertraline in phase 2 of the study (C).
Patients were excluded from the analyses if data were missing for them on the primary outcome for the second to last and last assessments (i.e., weeks 7 and 8 [phase 1] or weeks 12 and 16 [phase 2]), on more than one MRI sequence, or on pharmacological treatment for at least 2 weeks. We tested whether the exclusion of patients significantly affected the populations that were analyzed. We compared population characteristics before and after exclusion using Student’s t test for continuous variables and the chi-square test for categorical variables. The sociodemographic and clinical variables of the analyzed population are summarized in Table 1.
TABLE 1. Demographic, clinical, and outcome characteristics for pretreatment analysis subgroupsa
CharacteristicA. Phase 1 Sertraline (N=109)B. Phase 1 Placebo (N=120)C. Phase 2 Sertraline-Treated Placebo Nonresponders (N=58)
N%pN%pN%p
Female75690.9776630.8333571.00
Race/ethnicity
 African American23210.8917140.639161.00
 Asian661.000870.70231.00
 White72660.5686720.845721.00
 Other870.550870.694290.92
 Hispanic16150.3623190.8111190.78
Employed66610.4969571.0032550.83
MeanSDpMeanSDpMeanSDp
Age (years)38.114.00.8038.113.00.6739.913.00.93
Body mass index29.08.30.7328.86.90.4928.05.90.95
Years of education15.02.60.9115.22.40.8315.22.50.90
Age at symptom onset (years)15.85.70.6116.45.80.8716.45.40.97
Pretreatment HAM-D score18.74.40.7418.74.40.9818.73.80.69
SHAPS score28.49.80.9729.49.00.9532.37.90.80
MASQ score33.65.20.8433.16.00.9833.05.40.39
Posttreatment HAM-D score10.66.60.7912.37.50.4815.45.40.77
Change in HAM-D score−8.17.20.64−6.37.00.44−3.25.71.00
N%pN%pN%p
Outcome
 Remission46420.4036300.9228480.65
 Response55500.4141340.7637640.65
a
The p values express the difference before and after patient selection, using Student’s t test for continuous variables and the chi-square test for categorical variables. HAM-D=Hamilton Depression Rating Scale; MASQ=Mood and Anxiety Symptom Questionnaire; SHAPS=Snaith-Hamilton Pleasure Scale.

Treatment Outcome

HAM-D symptom severity outcomes were collected at weeks 8 and 16. The first primary clinical outcome is response, defined as a reduction ≥50% in HAM-D score after 8 weeks. The second is remission, defined as a HAM-D score ≤7 after 8 weeks.

Data Acquisition

Data were acquired before treatment and 1 week after treatment initiation (“early treatment”). The data consisted of clinical assessments and MRI neuroimaging. Clinical assessments consisted of sociodemographic, behavioral, and neuropsychological assessments. See Table S1 in the online supplement for details. Table S2 in the online supplement provides details on the MRI scanner and pulse sequence information by site. In brief, four MRI sequences were acquired: T1-weighted structural MRI, diffusion-weighted imaging, resting-state functional MRI (rs-fMRI), and perfusion MRI using arterial spin labeling (ASL). Diffusion-weighted imaging was acquired only at baseline. Although task-based fMRI was acquired, it was excluded from the present analysis because of the practical challenges it poses in clinical application.

MRI Preprocessing

All software programs used are listed in Table S3 in the online supplement. In short, T1-weighted scans were standardized using FreeSurfer (14) and segmented based on the Desikan-Killiany atlas (15) with FastSurfer (16). Volumes per anatomical region were computed using PyRadiomics (17). FreeSurfer segmentations underwent systematic quality control (QC) using MRIQC (18). Diffusion-weighted imaging data were preprocessed and analyzed using the JHU ICBM DTI-81 white matter tract atlas (19) and tract-based spatial statistics (20). QC was performed automatically using EDDY QC (21). For rs-fMRI, standardization and QC were performed using the fMRIPrep data preprocessing pipeline (22). Subsequently, dual-regression independent component analysis was performed using FSL (23). The resulting independent components were masked by the corresponding well-matched maps from the BrainMap resting-state networks (24) at Z score >3. ASL coregistration, analysis, and QC were performed using ExploreASL (25). Mean quantitative cerebral blood flow was extracted for regions of interest defined by the Talairach atlas (26). In addition to the various QC tools employed, all segmentations underwent visual inspection.

Definitions of Three Tiers of Predictors

For pretreatment prediction, pretreatment-acquired predictors were used, and for early-treatment prediction, we calculated relative reductions of predictors. Since MRI data were acquired only during phase 1, we did not perform early-treatment modeling when predicting phase 2 outcomes (i.e., week 16).
The clinical and MRI data contain many predictors. We limited the number of predictors using three predefined subsets (“tiers”) of predictors based on their level of scientific evidence. See Table S4 in the online supplement for a comprehensive list of these predictors.
Tier 1 predictors consisted of six MRI predictors for which level 1 scientific evidence exists, as indicated by meta-analyses and systematic reviews (summarized in reference 8). These predictors are the volumes of the left and right hippocampus, the mean and standard deviation of the frontal-limbic connectivity, and the mean and standard deviation of the quantitative cerebral blood flow in the anterior cingulate cortex (ACC). To these MRI predictors, the following clinical variables were added: age, sex, body mass index, chronicity, employment status, relative reduction in HAM-D score, and scores on the Mood and Anxiety Symptom Questionnaire, the Snaith-Hamilton Pleasure Scale, and the HAM-D. For tier 2 (see Table S4 in the online supplement), 54 predictors were added based on weaker (e.g., single study) evidence. Tier 3 consists of all 240 predictors without selection. For unimodal models, all predictors from other modalities were excluded, and the population was unaffected (Table 1, Figure 1). QC of produced predictors consisted of checking missing and outlier values.

Nested Cross-Validation

We followed stringent recommendations to reduce the risk of bias and overfitting and implemented a nested cross-validation framework with repeats. Imputation of scan and outcome data, site harmonization, scaling of data values, oversampling of outcomes, and hyperparameter optimization were embedded in the inner nest and tested in the outer cross-validation nest.
Train and test partitions for our primary hypothesis and single-modality models consisted of folds from population subgroup A (sertraline-treated patients). For our second hypothesis on sertraline specificity compared with placebo, we externally validated models that were trained on subgroup A on subgroups B (placebo-treated patients in phase 1) and C (sertraline-treated placebo nonresponders in phase 2). For internal validation, we performed 10-fold, 10-repeat, randomized K-fold cross-validation (RSKF) to improve our performance estimation. For external validation, we also applied 10 repeats since repeats affect the validation subset used during hyperparameter optimization. In addition, we performed a post hoc analysis where we compared RSKF with repeated leave-site-out cross-validation.

Data Preparation

Missing sequence-specific predictors were imputed using K-nearest neighbor imputation on data from the same missing sequence to prevent information leakage across MRI sequences. HAM-D outcomes were imputed using iterative multivariate conditional modeling on all available HAM-D scores. To mitigate confounding effects, we used an open-source harmonization method called ComBat (27). Covariates used in harmonization were age, age-squared, sex, brain volume, and site. See the methods section of the online supplement for more details. Response labels were balanced by random oversampling of the minority class without replacement.

Machine Learning Modeling

We trained XGBoost (28) extreme gradient boosting machine learning models implemented in Python with the area under the receiver operating characteristic curve (AUROC) as the criterion and a logistic objective function. Gradient boosting models have been shown to outperform other standard machine learning methods in classification studies related to MDD (29). In a post hoc analysis, we compared the results of this XGBoost classifier with support vector machine and logistic regression classifiers. See the methods section of the online supplement for details on the implementation of these classifiers. Hyperparameters (see Table S5 in the online supplement) were optimized using Bayesian hyperparameter optimization implemented for 100 epochs (30). To handle the large range of number of predictors provided to the model, feature selection was performed using chi-square-based feature selection (31) with the number of features selected optimized as a hyperparameter. (See Table S3 in the online supplement for details on the software used.)
Additionally, we report the most important predictors using impurity-based feature importance. Importance provides a score indicating how useful or valuable each feature was in constructing the boosted decision trees. For more information on the calculation of predictor importance, see the methods section of the online supplement.

Statistical Analysis

Primary predictive performance metrics are mean balanced accuracy (bAcc) and AUROC across repeats and folds. In addition, we report sensitivity and specificity. Balanced accuracy simplifies model evaluation independently of a priori outcome rates. We also computed the natural logarithm of the diagnostic odds ratio to allow comparison with a previous meta-analysis (10). For our primary hypothesis that our predictive models significantly outperform chance, we calculated the 95% confidence intervals for bAcc by chance, using one-tailed binomial tests using the number of unique test samples (32). Chance is defined as the a priori response rate, the best alternative method. As a supplementary (post hoc) analysis, we predicted continuous clinical outcome measures directly; see the methods section of the online supplement for more details.
To test our second hypothesis on treatment specificity, we tested whether external validation performance in subgroups B (placebo-treated) and C (sertraline-treated placebo nonresponders) was significantly lower than on internal validation in subgroup A (sertraline-treated) by performing a one-sided dependent t test between subgroup B and the independent test set A, and between subgroup C and test set A. The t test samples consisted of the mean performance estimates for all modeling configurations, that is, pretreatment and early-treatment prediction of remission and response. We excluded configurations that did not perform significantly better than chance from this analysis to avoid regression toward the no-information rate.
Finally, we tested whether multimodal models outperformed unimodal models. We tested the null hypothesis that their mean performances were identical. We used one-tailed sign tests between the unimodal and the multimodal AUROC score pairs and a p value of 0.05 as the significance threshold (33).

Results

Patient Selection

A total of 229 patients were included in the analysis; their mean age was 38.1 years (SD=13.3), and 65.9% were female (Table 1, Figure 1). Details on causes for missing data are provided in Table S6 in the online supplement. Also, excluding patients with missing primary outcomes, primarily owing to lack of follow-up, did not significantly affect population characteristics. Treatment adherence in the remaining cohort was generally high, since among patients included there were none who stopped medication for more than a week (see Table S6 in the online supplement).

Multimodal Modeling

An overview of mean cross-validation results is presented in Table 2. All tier 1 models achieved significantly better performance than chance, in addition to one tier 2–based model and two tier 3–based models. Additional metrics are provided in Table S7 in the online supplement. For the prediction of treatment response with sertraline, best model performance was achieved in the early-treatment prediction of response (bAcc=68% [SD=2.4], AUROC=0.73 [SD=0.03], sensitivity=70% [SD=3.3], specificity=70% [SD=3.3], ln[DOR]=2.46 [SD=1.10], respectively) (Figure 2, panel 1, and Table 2). This performance is significantly better than chance (p<0.001). On average, early-treatment performance (bAcc=66.5% [SD=4.2], AUROC=0.72 [SD=0.11]) was better than pretreatment performance (bAcc=63% [SD=4.2], AUROC=0.72 [SD=4.7]), but the difference was not significant (p=0.06). The performance was higher for models that used predictors with a higher level of scientific evidence; all the best models used tier 1 predictors. Results of the post hoc analyses using leave-site-out cross-validation (see Table S8 in the online supplement), the comparison of machine learning classifiers (see Table S9 in the online supplement), and prediction of continuous clinical outcomes (see Table S10 and Figure S1) are provided in the online supplement. The mean absolute error on tier 1–based regression models achieved a mean absolute error of 5 points on the HAM-D scale. Regression models performed worse on the binary classification task than the corresponding classifiers presented in the main work.
TABLE 2. Main model performance metrics for all configurationsa
Predictor Tier1. Tested on A. Phase 1 Sertraline2. Tested on B. Phase 1 Placebo3. Tested on C. Phase 2 Sertraline-Treated Placebo Nonresponders4. AUC for Single Modalities
AUCbAcc (%)SensSpecAUCbAcc (%)SensSpecAUCbAcc (%)SensSpecClinT1wrs-fMRIASLDWI
Pretreatment prediction of remission
Tier 10.68b640.650.670.59b590.580.610.68b640.630.630.59b0.530.460.65b0.46
Tier 20.59b560.570.550.60b590.590.600.67b640.650.620.60b0.450.470.60b0.48
Tier 30.58b570.580.590.58b560.550.570.58560.560.560.48b0.540.490.58b0.48
Pretreatment prediction of response
Tier 10.71b650.660.660.57b570.580.560.56520.540.520.54b0.540.500.63b0.48
Tier 20.61b580.580.610.58b550.560.550.52520.520.510.52b0.460.500.60b0.50
Tier 30.65b610.600.650.58b530.540.530.50460.410.470.53b0.590.480.60b0.50
Early-treatment prediction of remission
Tier 10.69b620.640.650.62b570.570.580.58b0.530.500.62b
Tier 20.59b560.570.560.62b570.570.580.64b0.510.480.51b
Tier 30.62b590.600.580.66b600.590.620.62b0.460.610.53b
Early-treatment prediction of response
Tier 10.73b680.700.700.60b570.590.560.65b0.550.500.56b
Tier 20.63b600.610.610.63b580.600.570.62b0.450.450.57b
Tier 30.60b580.580.600.61b570.570.560.58b0.490.580.56b
a
AUC=area under the receiver operating characteristic curve; bAcc=balanced accuracy; sens=sensitivity; spec=specificity; clin=clinical assessment data; T1w=T1-weighted imaging; rs-fMRI=resting-state functional MRI; ASL=arterial spin labeling; DWI=diffusion-weighted imaging.
b
Models with an area under the receiver operating characteristic curve significantly above chance with a p value <0.05.
FIGURE 2. Receiver operating characteristic curves of the best-performing configuration of each analysisa
aPanel A shows the mean receiver operating characteristic curve (ROC) over 10 folds from subgroup A for early-treatment response prediction using tier 1 predictors; panel B shows early-treatment prediction of remission on subgroup B using tier 3 predictors; and panel C shows pretreatment prediction of remission on subgroup C using tier 1 predictors. AUC=area under the curve.
The most important pretreatment predictors were ASL predictors. The median importance of the mean and standard deviation of the quantitative cerebral blood flow in the ACC is 0.13. Note that importance sums to 1. The third most important predictor was the anhedonic depression score on the Mood and Anxiety Symptom Questionnaire (0.11). However, at early treatment, we saw a shift from ASL predictors to clinical predictors. The most important predictors were reduction in HAM-D score (0.17), HAM-D score at week 1, and anhedonic depression score on the Mood and Anxiety Symptom Questionnaire at baseline (0.12). All predictors mentioned were consistently selected through hyperparameter optimization. For more details, see Figure S2 and Table S11 in the online supplement.

Multimodal Model Specificity for Sertraline Treatment

For model specificity to treatment, performance was significantly reduced (p=0.01) when tested on placebo-treated patients in subgroup B (Table 2). However, half the models still achieved performance significantly better than chance compared with the primary analysis, being tier 1- and tier 2–based pretreatment and early-treatment prediction of remission and tier 2–based early-treatment prediction of response. An exception to this lower performance was the early-treatment prediction of remission (bAcc=64%, AUROC=0.66) in subgroup B using tier 3 predictors (Figure 2, panel 2), which outperformed the same model in subgroup A (bAcc=59% [SD=2.7], AUROC=0.62 [SD=0.05]).
When tested on the sertraline-treated placebo nonresponders of subgroup C, mean predictive performance was not lower than internal validation (p=0.06). The performance of predicting remission in subgroup C (bAcc=64%, AUROC=0.68 (Figure 2, panel 3) using tier 1 predictors was similar to that of internal validation in subgroup A (bAcc=64%, AUROC=0.68) (Figure 2 panel 1), and both tier 1 and 2 models performed significantly better than chance. However, performance on predicting response in subgroup C using tier 1 predictors (bAcc=56%, AUROC=0.56) was worse compared with subgroup A (bAcc=65% [SD=4.1], AUROC=0.71 [SD=0.06]).

Unimodal Modeling

The mean performance of multimodal models significantly outperformed all unimodal models except early-treatment and early-treatment clinical assessment-based prediction of remission (Figure 3). Statistical outcomes are listed in Table S12 in the online supplement. The 11 unimodal models that performed significantly better than chance consisted of six pretreatment models using ASL and five early-treatment models using clinical assessment data (Table 2).
FIGURE 3. Box plots of best cross-validation performance of multimodal versus unimodal modelsa
aBoxes show two interquartile ranges with a median centerline. An asterisk marks unimodal cross-validation results that are significantly (p<0.05) worse than their multimodal counterpart. See Table S12 in the online supplement for statistical outcomes. ASL=arterial spin labeling; AUROC=area under the receiver operating characteristic curve; DWI=diffusion-weighted imaging; rs-fMRI=resting-state functional MRI.

Discussion

We demonstrated that response in MDD can be predicted using pretreatment and early-treatment multimodal MRI and clinical data. Validation on the placebo arm and sertraline-treated placebo nonresponders suggested specificity for sertraline treatment compared with placebo treatment. Our results indicated that predictors with strong scientific evidence are the primary drivers of model performance. Finally, multimodal prediction outperformed most unimodal approaches, and models using ASL predictors showed the best unimodal predictions, even with pretreatment data only.

Multimodal Modeling

Compared with other multimodal studies, our results confirm the positive conclusions of Patel et al. (34) and Leaver et al. (35). These studies’ small sample sizes of 44 and 19 patients, respectively, raise the possibility that their findings are affected by performance bias (36). More recently, Sajjadian et al. (11) presented an approach using data from the Canadian Biomarker Integration Network in Depression (CAN-BIND-1) study. Compared with that study, our results are an improvement both in terms of bAcc and AUROC on corresponding analyses (pretreatment prediction of response: bAcc, 65% vs. 60%; AUROC, 0.71 vs. 0.60; early-treatment prediction of response: bAcc, 68% vs. 66%; AUROC, 0.73 vs. 0.70). This improvement is larger for pretreatment prediction than for early-treatment prediction. Since CAN-BIND-1 follow-up clinical data were acquired at week 2 instead of week 1 as in EMBARC, we suspect that the CAN-BIND-1 clinical data might present a larger early-treatment effect. Since the two approaches are methodologically similar, another explanation for our improved results is the availability of ASL in EMBARC but not in CAN-BIND-1. Our unimodal results substantiate this hypothesis, indicating significant predictive power for ASL alone.
We performed three post hoc analyses: First, our findings are consistent with leave-site-out cross-validation (see Table S8 in the online supplement), suggesting that site effect is limited. Second, comparing alternative classifiers to the XGBoost classifier shows that these are also capable of predicting treatment response, albeit at lower performance levels (see Table S9 in the online supplement). Lastly, we found that regression models can predict continuous outcome scores (see Table S10 and Figure S1 in the online supplement).
When compared with the predictive performance reported by two recent meta-analyses of treatment response prediction based on MRI for pharmacotherapy, our approach outperformed the weighted estimate of the natural logarithm of the diagnostic odds ratio (2.46 [SD=1.10] and 2.11 [SD=0.56], respectively) (10), as well as the mean bAcc for adequate-quality studies (68% [SD=10] and 63% [SD=7], respectively) (36).
Several considerations should be borne in mind when qualitatively evaluating the utility of the model. First, it should be noted that with the accuracy measure used in this work (balanced accuracy), the actual accuracy will be higher when the response rate is greater than 50%. An AUROC >0.7 can be considered good, depending on the use. In our case, our model outperforms the current trial-and-error standard used in clinical practice. Thus, the benefits of utilizing our treatment planning support tool, which can predict treatment efficacy early on, might quickly outweigh the costs of MRI scanning. Additionally, our work outperforms studies of similar scope. However, the clinical effectiveness of treatment decision support tools should be determined in independent prospective randomized clinical trials.
In clinical practice, it might also be desirable to increase the detection rate of nonresponders at the cost of decreased sensitivity (indicating nonresponse in final responders). This would allow improved identification of true nonresponders to switch to another treatment early at a higher, acceptable rate of false “nonresponder” assignments among patients who might respond to a second selective serotonin reuptake inhibitor (SSRI) or serotonin-norepinephrine reuptake inhibitor (37). Finally, the lower bound at which performance predictive modeling using MRI balances burden and cost should be investigated.

Multimodal Model Specificity for Sertraline Treatment

The results for the sertraline-treated placebo nonresponders (subgroup C) were not significantly different from the sertraline-treated training set of subgroup A. In contrast, performance in the placebo-treated subgroup B did decrease significantly. This contrast suggests some specificity for predicting sertraline compared with placebo treatment response. Nevertheless, external validation in subgroup B achieved performance significantly better than chance, indicating some overlap. One explanation suggested by the literature is that, in addition to the pharmacotherapeutic effect, sertraline treatment is partially driven by a placebo response (38). Further studies are required to establish the specificity of our approach for other antidepressants.

Unimodal Versus Multimodal Modeling

Our results indicate consistently significantly higher performance in multimodal compared with unimodal models. This supports the claim that integrating neuroimaging data with clinical data can improve treatment response prediction. However, models of two modalities did perform significantly better than chance: pretreatment models using ASL only and early-treatment models using clinical assessment data only. These two data types are primary candidates for a lean treatment response model. As an exploratory analysis, we report on the performance of these models in Table S13 in the online supplement. Nevertheless, our findings corroborate the call for multimodal data integration to improve predictive performance (9).
Although differences in study design limit comparisons with other studies, the performance of unimodal models was similar to that of previous unimodal studies (this vs. other). This confirms results achieved by previous studies and is a sign that our multimodal results may be similar in other study populations. Bartlett et al. (39) predicted remission based on T1-weighted MRI (bAcc, 53% vs. 51%, and AUROC, 0.55 vs. 0.59, respectively). Korgaonkar et al. (40) predicted remission using baseline diffusion-weighted imaging (bAcc, 52% vs. 54%). Finally, Chekroud et al. (41) used early-treatment clinical variables for predictions of remission (bAcc, 65% vs. 63%, and AUROC, 0.69 vs. 0.70, respectively).

Interpretability

Balancing prediction performance with interpretability is an ongoing challenge in neuroimaging-based modeling. However, given that even the best-performing neuroimaging-based models show modest prediction performance, more emphasis on interpretability may be warranted (42). Therefore, we performed a post hoc exploration of feature importance, as shown in Figure S2 and Table S11 in the online supplement. Early-treatment response and perfusion in the ACC consistently contributed to our model’s performance. ACC perfusion is a well-replicated biomarker for treatment response (8), and this finding aligns with previous work in MDD. A recent meta-analysis found that the ACC plays a crucial role in treatment outcome because of its role as a hub in the interplay between the ventral emotion-generating and dorsal cognitive-controlling systems. Here, SSRIs are thought to play a role in improving this ventral-dorsal control (43). After 1 week, we observed a high importance of symptom severity reduction, a known predictor (44). Unexpectedly, the hippocampal volume—the most replicated finding concerning treatment response in MDD (8)—did not consistently contribute. A possible explanation is that predictors were selected individually, while a strong effect of left lateralization of hippocampal volume in treatment prediction has been noted (45). Similarly, the connectivity in the default mode network measured with rs-fMRI did not contribute consistently to our models. For more information on predictor importance, see Figure S2 and Table S11 in the online supplement.

Strengths and Limitations

The design of the EMBARC trial allowed us to increase the generalizability of our results in three ways. First, EMBARC is the largest multimodal neuroimaging MDD data set available, and the sample size has been shown to reduce bias in performance estimation (9, 36). Second, data were acquired at four sites using 3-T scanners from different manufacturers. We preprocessed and harmonized the data using standard software packages to increase generalizability to new populations (46) and tested our results using leave-site-out cross-validation. Third, we performed validation in a separate study arm, instead of just internal cross-validation as in most studies (10). Performance was not significantly reduced in the separate set of sertraline-treated patients, which is a positive indicator of the generalizability of our method. An additional strength is that the hypotheses and methods in this work were preregistered prior to analysis. Preregistration avoids overestimation of performance due to overfitting, which happens when algorithms are tailored to training data after data analysis. Although preregistration of tier evidence (8) inhibits the exploration of other combinations of predictors, it increases the reproducibility and validity of our results since the prespecified predictors (tier 1) were backed by strong scientific evidence (systematic reviews or meta-analyses).
While the study design provides valuable strengths, several limitations should be acknowledged. First, our results are limited to sertraline treatment from data acquired in the EMBARC study. In the EMBARC study, information about response to previous antidepressant treatment was not available, which might be valuable to add to future predictive tools (47). Thus, further validation of our approach is required on even larger collections of external data, for different antidepressants, and in populations with more clinical heterogeneity than in a randomized clinical trial. Until then, our results should be interpreted with caution. To our knowledge, no data set exists that has collected the same multimodal data at the same time points as the EMBARC trial that would allow such external validation. Second, although excluding patients did not significantly affect population characteristics, excluding patients who did not complete the study’s first phase may have favored prediction results in the direction of adherent participants. Third, the EMBARC study’s design allowed us to validate our models on the same patients treated with placebo (subgroup B) and sertraline (subgroup C). However, subgroup C lacked patients from subgroup B who had responded to placebo treatment in the first phase of the study. This selection can be expected to incur a bias in our results. Since response is known to be partially driven by placebo response (38), lowered treatment response and decreased treatment response prediction performance in subgroup C could have been expected. However, the response rate in subgroup C was not negatively affected by this selection compared with subgroup A (Table 1). Our results also show that performance in subgroup C was not significantly reduced compared with subgroup A. Thus, we can assume that this type of selection bias did not affect our conclusions. Finally, although task-based fMRI might provide promising information to add to multimodal modeling (4850) and likewise provide important insights into context-dependent neural mechanisms, we excluded this option, given the difficulty of their scalability and replicability in clinical practice (51). Therefore, we could not assess its potential benefits in treatment response prediction (8).
We found no significant difference in performance between pretreatment and early-treatment prediction. If externally validated, early-treatment response prediction would likely not require a second session of MRI scanning, reducing cost and lowering patient burden. Because our results show a performance drop in unimodal ASL models at early treatment but not at pretreatment, and we decided a priori to use relative changes at early treatment, which are sensitive to physiological variability, we suspect that this choice was suboptimal for ASL predictors. To overcome this limitation, we suggest combining absolute ASL predictors with early-treatment clinical predictors to improve performance. Other improvement options include integrating predictors from other sources and utilizing novel analytic methods, such as normative modeling (52).
In summary, our findings on a multimodal machine-learning-based method applied to data from the EMBARC trial show that pretreatment and early-treatment prediction of sertraline treatment response in MDD patients is feasible using brain MRI and clinical data and significantly outperforms chance and most unimodal models. Our results also suggest the specificity of our models for sertraline compared with placebo treatment. We found that ASL was the best unimodal predictor. With additional external validation, these findings will contribute toward the use of predictive modeling in individualizing clinical sertraline treatment of patients with MDD.

Supplementary Material

File (ajp_20230206_correction01.pdf)
File (appi.ajp.20230206.ds001.pdf)

References

1.
Bromet E, Andrade LH, Hwang I, et al: Cross-national epidemiology of DSM-IV major depressive episode. BMC Med 2011; 9:90
2.
Kraemer HC: Messages for clinicians: moderators and mediators of treatment outcome in randomized clinical trials. Am J Psychiatry 2016; 173:672–679
3.
Trivedi MH, Rush AJ, Wisniewski SR, et al: Evaluation of outcomes with citalopram for depression using measurement-based care in STAR*D: implications for clinical practice. Am J Psychiatry 2006; 163:28–40
4.
Al-Harbi KS: Treatment-resistant depression: therapeutic trends, challenges, and future directions. Patient Prefer Adherence 2012; 6:369–388
5.
Greenberg PE, Fournier AA, Sisitsky T, et al: The economic burden of adults with major depressive disorder in the United States (2005 and 2010). J Clin Psychiatry 2015; 76:155–162
6.
Phillips ML, Chase HW, Sheline YI, et al: Identifying predictors, moderators, and mediators of antidepressant response in major depressive disorder: neuroimaging approaches. Am J Psychiatry 2015; 172:124–138
7.
Lee Y, Ragguett RM, Mansur RB, et al: Applications of machine learning algorithms to predict therapeutic outcomes in depression: a meta-analysis and systematic review. J Affect Disord 2018; 241:519–532
8.
Schrantee A, Ruhe HG, Reneman L: Psychoradiological biomarkers for psychopharmaceutical effects. Neuroimaging Clin N Am 2020; 30:53–63
9.
Schmaal L: The search for clinically useful neuroimaging markers of depression: a worthwhile pursuit or a futile quest? JAMA Psychiatry 2022; 79:845–846
10.
Cohen SE, Zantvoord JB, Wezenberg BN, et al: Magnetic resonance imaging for individual prediction of treatment response in major depressive disorder: a systematic review and meta-analysis. Transl Psychiatry 2021; 11:168
11.
Sajjadian M, Uher R, Ho K, et al: Prediction of depression treatment outcome from multimodal data: a CAN-BIND-1 report. Psychol Med 2022; 53:5374–5384
12.
Poirot MG, Ruhe HG, Maximov II, et al: Pretreatment and early-treatment prediction of SSRI treatment response in major depressive disorder: a multimodal MRI approach on the EMBARC dataset. Open Science Framework Registry, 2022. (https://doi.org/10.17605/OSF.IO/79CWY)
13.
Trivedi MH, McGrath PJ, Fava M, et al: Establishing Moderators and Biosignatures of Antidepressant Response in Clinical Care (EMBARC): rationale and design. J Psychiatr Res 2016; 78:11–23
14.
Fischl B, Salat DH, Busa E, et al: Whole brain segmentation: automated labeling of neuroanatomical structures in the human brain. Neuron 2002; 33:341–355
15.
Desikan RS, Ségonne F, Fischl B, et al: An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. Neuroimage 2006; 31:968–980
16.
Henschel L, Conjeti S, Estrada S, et al: FastSurfer: a fast and accurate deep learning based neuroimaging pipeline. Neuroimage 2020; 219:117012
17.
van Griethuysen JJM, Fedorov A, Parmar C, et al: Computational radiomics system to decode the radiographic phenotype. Cancer Res 2017; 77:e104–e107
18.
Esteban O, Birman D, Schaer M, et al: MRIQC: advancing the automatic prediction of image quality in MRI from unseen sites. PLoS One 2017; 12:e0184661
19.
Mori S, Oishi K, Jiang H, et al: Stereotaxic white matter atlas based on diffusion tensor imaging in an ICBM template. Neuroimage 2008; 40:570–582
20.
Yeh P-H, Simpson K, Durazzo TC, et al: Tract-based spatial statistics (TBSS) of diffusion tensor imaging data in alcohol dependence: abnormalities of the motivational neurocircuitry. Psychiatry Res 2009; 173:22–30
21.
Bastiani M, Cottaar M, Fitzgibbon SP, et al.: Automated quality control for within and between studies diffusion MRI data using a non-parametric framework for movement and distortion correction. Neuroimage 2019; 184:801–812
22.
Esteban O, Markiewicz CJ, Blair RW, et al: fMRIPrep: a robust preprocessing pipeline for functional MRI. Nat Methods 2019; 16:111–116
23.
Beckmann CF, Mackay CE, Filippini N, et al: Group comparison of resting-state fMRI data using multi-subject ICA and dual regression. Neuroimage 2009; 47:S148
24.
Smith SM, Fox PT, Miller KL, et al: Correspondence of the brain’s functional architecture during activation and rest. Proc Natl Acad Sci U S A 2009; 106:13040–13045
25.
Mutsaerts HJMM, Petr J, Groot P, et al: ExploreASL: an image processing pipeline for multi-center ASL perfusion MRI studies. Neuroimage 2020; 219:117031
26.
Talairach J, Tournoux P: Co-Planar Stereotaxic Atlas of the Human Brain: 3-Dimensional Proportional System: An Approach to Cerebral Imaging. New York, Thieme Medical Publishers, 1988
27.
Johnson WE, Li C, Rabinovic A: Adjusting batch effects in microarray expression data using empirical Bayes methods. Biostatistics 2007; 8:118–127
28.
Chen T, Guestrin C: XGBoost: a scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016, pp 785–794
29.
Gao S, Calhoun VD, Sui J: Machine learning in major depression: from classification to treatment outcome prediction. CNS Neurosci Ther 2018; 24:1037–1052
30.
Bergstra J, Yamins D, Cox D: Making a science of model search: hyperparameter optimization in hundreds of dimensions for vision architectures. 30th International Conference on Machine Learning, 2013, pp 115–123
31.
Pedregosa F, Varoquaux G, Gramfort A, et al: Scikit-learn: machine learning in Python. J Mach Learn Res 2011; 12:2825–2830
32.
Combrisson E, Jerbi K: Exceeding chance level by chance: the caveat of theoretical chance levels in brain signal classification and statistical assessment of decoding accuracy. J Neurosci Methods 2015; 250:126–136
33.
Dixon WJ, Mood AM: The statistical sign test. J Am Stat Assoc 1946; 41:557–566
34.
Patel MJ, Andreescu C, Price JC, et al: Machine learning approaches for integrating clinical and imaging features in late-life depression classification and response prediction. Int J Geriatr Psychiatry 2015; 30:1056–1067
35.
Leaver AM, Wade B, Vasavada M, et al: Fronto-temporal connectivity predicts ECT outcome in major depression. Front Psychiatry 2018; 9:92
36.
Sajjadian M, Lam RW, Milev R, et al: Machine learning in the prediction of depression treatment outcomes: a systematic review and meta-analysis. Psychol Med 2021; 51:2742–2751
37.
Tadić A, Wachtlin D, Berger M, et al: Randomized controlled study of early medication change for non-improvers to antidepressant therapy in major depression: the EMC trial. Eur Neuropsychopharmacol 2016; 26:705–716
38.
Burke MJ, Romanella SM, Mencarelli L, et al: Placebo effects and neuromodulation for depression: a meta-analysis and evaluation of shared mechanisms. Mol Psychiatry 2022; 27:1658–1666
39.
Bartlett EA, DeLorenzo C, Sharma P, et al: Pretreatment and early-treatment cortical thickness is associated with SSRI treatment response in major depressive disorder. Neuropsychopharmacology 2018; 43:2221–2230
40.
Korgaonkar MS, Williams LM, Song YJ, et al: Diffusion tensor imaging predictors of treatment outcomes in major depressive disorder. Br J Psychiatry 2014; 205:321–328
41.
Chekroud AM, Zotti RJ, Shehzad Z, et al: Cross-trial prediction of treatment outcome in depression: a machine learning approach. Lancet Psychiatry 2016; 3:243–250
42.
Tejavibulya L, Rolison M, Gao S, et al: Predicting the future of neuroimaging predictive models in mental health. Mol Psychiatry 2022; 27:3129–3137
43.
Pizzagalli DA: Frontocingulate dysfunction in depression: toward biomarkers of treatment response. Neuropsychopharmacology 2011; 36:183–206
44.
Szegedi A, Jansen WT, van Willigenburg APP, et al: Early improvement in the first 2 weeks as a predictor of treatment outcome in patients with major depressive disorder: a meta-analysis including 6562 patients. J Clin Psychiatry 2009; 70:344–353
45.
Hu X, Zhang L, Hu X, et al: Abnormal hippocampal subfields may be potential predictors of worse early response to antidepressant treatment in drug-naïve patients with major depressive disorder. J Magn Reson Imaging 2019; 49:1760–1768
46.
Orlhac F, Eertink JJ, Cottereau AS, et al: A guide to ComBat harmonization of imaging biomarkers in multicenter studies. J Nucl Med 2022; 63:172–179
47.
Paul R, Andlauer TFM, Czamara D, et al: Treatment response classes in major depressive disorder identified by model-based clustering and validated by clinical prediction models. Transl Psychiatry 2019; 9:187
48.
Nguyen KP, Fatt CC, Treacher A, et al: Predicting response to the antidepressant bupropion using pretreatment fMRI. Predict Intell Med 2019; 11843:53–62
49.
Nguyen KP, Chin Fatt C, Treacher A, et al: Patterns of pretreatment reward task brain activation predict individual antidepressant response: key results from the EMBARC randomized clinical trial. Biol Psychiatry 2022; 91:550–560
50.
Chase HW, Fournier JC, Greenberg T, et al: Accounting for dynamic fluctuations across time when examining fMRI test-retest reliability: analysis of a reward paradigm in the EMBARC study. PLoS One 2015; 10:e0126326
51.
Klomp A, van Wingen GA, de Ruiter MB, et al: Test-retest reliability of task-related pharmacological MRI with a single-dose oral citalopram challenge. Neuroimage 2013; 75:108–116
52.
Rutherford S, Barkema P, Tso IF, et al: Evidence for embracing normative modeling. eLife 2023; 12:e85082

Information & Authors

Information

Published In

Go to American Journal of Psychiatry
Go to American Journal of Psychiatry
American Journal of Psychiatry
Pages: 223 - 233
PubMed: 38321916

History

Received: 15 March 2023
Revision received: 21 July 2023
Revision received: 14 September 2023
Accepted: 12 October 2023
Published online: 7 February 2024
Published in print: March 01, 2024

Keywords

  1. Major Depressive Disorder
  2. Machine Learning
  3. Antidepressants
  4. SSRIs

Authors

Details

Maarten G. Poirot, M.S.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Henricus G. Ruhe, M.D., Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Henk-Jan M.M. Mutsaerts, M.D., Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Ivan I. Maximov, Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Inge R. Groote, M.D., Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Atle Bjørnerud, Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Henk A. Marquering, Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Liesbeth Reneman, M.D., Ph.D. [email protected]
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).
Matthan W.A. Caan, Ph.D.
Department of Radiology and Nuclear Medicine (Poirot, Ruhe, Marquering, Reneman) and Department of Biomedical Engineering and Physics (Poirot, Marquering, Reneman, Caan), Amsterdam UMC Location AMC, University of Amsterdam, Amsterdam; Brain Imaging, Amsterdam Neuroscience, Amsterdam (Poirot, Mutsaerts, Reneman, Caan); Department of Psychiatry, Radboudumc, Nijmegen, the Netherlands (Ruhe); Donders Institute for Brain, Cognition, and Behavior, Radboud University, Nijmegen, the Netherlands (Ruhe); Department of Radiology and Nuclear Medicine, Amsterdam UMC location, Vrije Universiteit Amsterdam, Amsterdam (Mutsaerts); Department of Health and Functioning, Western Norway University of Applied Sciences, Bergen (Maximov, Bjørnerud); Division of Radiology, Vestfold Hospital Trust, Tønsberg, Norway (Groote, Caan); Computational Radiology and Artificial Intelligence, Division of Radiology and Nuclear Medicine, Oslo University Hospital, Oslo (Groote, Bjørnerud); Department of Psychology, University of Oslo, Oslo (Bjørnerud).

Notes

Send correspondence to Dr. Reneman ([email protected]).

Author Contributions

Drs. Reneman and Caan contributed equally as senior authors.

Funding Information

Supported by the Eurostars funding program (reference number 113351). Dr. Mutsaerts is supported by the Dutch Heart Foundation (03-004-2020-T049), by the Eurostars-2 joint program with co-funding from the European Union Horizon 2020 research and innovation program (ASPIRE E!113701), provided by the Netherlands Enterprise Agency (RvO), and by the EU Joint Program for Neurodegenerative Disease Research, provided by the Netherlands Organization for Health Research and Development and Alzheimer Nederland (DEBBIE JPND2020-568-106). The EMBARC study was supported by NIMH under grants U01MH092221 and U01MH092250. The EMBARC study was also supported in part by Valeant Pharmaceuticals, which donated extended-release bupropion, used in the second phase of the EMBARC study (not reported here).Dr. Ruhe has received research grants from the Dutch Ministry of Health, EU Horizon 2020, HersenStichting, and ZonMW and an unrestricted educational grant from Janssen, and he has received speaking honoraria from Benecke, Janssen, Lundbeck, Prelum, and Wyeth. Dr. Marquering is a co-founder of and shareholder in InSteps, Nicolab, and TrianecT. Dr. Caan is a shareholder in Nicolab. The other authors report no financial relationships with commercial interests.

Metrics & Citations

Metrics

Citations

Export Citations

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

For more information or tips please see 'Downloading to a citation manager' in the Help menu.

Format
Citation style
Style
Copy to clipboard

View Options

View options

PDF/EPUB

View PDF/EPUB

Login options

Already a subscriber? Access your subscription through your login credentials or your institution for full access to this article.

Personal login Institutional Login Open Athens login
Purchase Options

Purchase this article to access the full text.

PPV Articles - American Journal of Psychiatry

PPV Articles - American Journal of Psychiatry

Not a subscriber?

Subscribe Now / Learn More

PsychiatryOnline subscription options offer access to the DSM-5-TR® library, books, journals, CME, and patient resources. This all-in-one virtual library provides psychiatrists and mental health professionals with key resources for diagnosis, treatment, research, and professional development.

Need more help? PsychiatryOnline Customer Service may be reached by emailing [email protected] or by calling 800-368-5777 (in the U.S.) or 703-907-7322 (outside the U.S.).

Figures

Tables

Media

Share

Share

Share article link

Share