Hostname: page-component-89b8bd64d-9prln Total loading time: 0 Render date: 2026-05-12T22:35:11.477Z Has data issue: false hasContentIssue false

Select or adjust? How information from early treatment stages boosts the prediction of non-response in internet-based depression treatment

Published online by Cambridge University Press:  13 December 2023

Leona Hammelrath*
Affiliation:
Department of Education and Psychology, Freie Universität Berlin, Berlin, Germany
Kevin Hilbert
Affiliation:
Department of Psychology, Health and Medical University Erfurt, Erfurt, Germany
Manuel Heinrich
Affiliation:
Department of Education and Psychology, Freie Universität Berlin, Berlin, Germany
Pavle Zagorscak
Affiliation:
Department of Education and Psychology, Freie Universität Berlin, Berlin, Germany
Christine Knaevelsrud
Affiliation:
Department of Education and Psychology, Freie Universität Berlin, Berlin, Germany
*
Corresponding author: Leona Hammelrath; Email: leona.hammelrath@fu-berlin.de
Rights & Permissions [Opens in a new window]

Abstract

Background

Internet-based interventions produce comparable effectiveness rates as face-to-face therapy in treating depression. Still, more than half of patients do not respond to treatment. Machine learning (ML) methods could help to overcome these low response rates by predicting therapy outcomes on an individual level and tailoring treatment accordingly. Few studies implemented ML algorithms in internet-based depression treatment using baseline self-report data, but differing results hinder inferences on clinical practicability. This work compares algorithms using features gathered at baseline or early in treatment in their capability to predict non-response to a 6-week online program targeting depression.

Methods

Our training and test sample encompassed 1270 and 318 individuals, respectively. We trained random forest algorithms on self-report and process features gathered at baseline and after 2 weeks of treatment. Non-responders were defined as participants not fulfilling the criteria for reliable and clinically significant change on PHQ-9 post-treatment. Our benchmark models were logistic regressions trained on baseline PHQ-9 sum or PHQ-9 early change, using 100 iterations of randomly sampled 80/20 train-test-splits.

Results

Best performances were reached by our models involving early treatment characteristics (recall: 0.75–0.76; AUC: 0.71–0.77). Therapeutic alliance and early symptom change constituted the most important predictors. Models trained on baseline data were not significantly better than our benchmark.

Conclusions

Fair accuracies were only attainable by involving information from early treatment stages. In-treatment adaptation, instead of a priori selection, might constitute a more feasible approach for improving response when relying on easily accessible self-report features. Implementation trials are needed to determine clinical usefulness.

Information

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - SA
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike licence (http://creativecommons.org/licenses/by-nc-sa/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the same Creative Commons licence is used to distribute the re-used or adapted article and the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use.
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press
Figure 0

Table 1. Patient summary characteristics stratified by treatment outcome

Figure 1

Figure 1. Comparison of model performance in identifying non-response.Notes: The dashed line indicates chance level. RF, random forest; base, pre-treatment features only; early, features from the beginning of week 2 were incorporated; LogReg base, logistic regression using the baseline PHQ sum; LogReg EC, logistic regression using PHQ difference from baseline to week 2 (early change)

Figure 2

Table 2. Outcomes by model type averaged across 100 iterations

Figure 3

Figure 2. ROC curves for the random forest model involving early treatment features, hyperparameter tuning and automatic feature selection.Notes: The dashed line indicates chance level. The bold line indicates the averaged ROC across 100 iterations for the random forest model. The bold dotted line indicated the averages ROC across 100 iterations for the benchmark model trained on PHQ early-change. RF, random forest; AUC, area under the curve; SD, standard deviation.

Figure 4

Figure 3. The 10 most important features for the random forest involving early treatment features, hyperparameter tuning and feature selection.Notes: Importance is computed by the Gini impurity index averaged across iterations. The numbers above the bars indicate the amount of rounds the feature has been selected by automatic feature selection. There were 100 iterations in total.

Supplementary material: File

Hammelrath et al. supplementary material

Hammelrath et al. supplementary material
Download Hammelrath et al. supplementary material(File)
File 29.2 KB