Hostname: page-component-77f85d65b8-2tv5m Total loading time: 0 Render date: 2026-04-13T06:14:51.588Z Has data issue: false hasContentIssue false

Evaluating the patient journey through integrated mental health services using routinely collected data: utility of the DIALOG patient-reported outcome and experience measure

Published online by Cambridge University Press:  01 April 2026

Stuart G. Spicer*
Affiliation:
Community & Primary Care Research Centre, University of Plymouth, Plymouth, UK
Rahul Bhattacharya
Affiliation:
East London NHS Foundation Trust, Tower Hamlets Directorate, London, UK Warwick Medical School, University of Warwick, Coventry, UK
Katelyn Smalley
Affiliation:
Community & Primary Care Research Centre, University of Plymouth, Plymouth, UK
Akshith Shetty
Affiliation:
North East London NHS Foundation Trust, London, UK
Paul Sharpe
Affiliation:
Community & Primary Care Research Centre, University of Plymouth, Plymouth, UK
Richard Byng
Affiliation:
Community & Primary Care Research Centre, University of Plymouth, Plymouth, UK
*
Correspondence to Stuart G. Spicer (stuart.spicer@plymouth.ac.uk)
Rights & Permissions [Opens in a new window]

Abstract

Aims and method

DIALOG is a patient-reported outcome and experience measure. We analysed anonymised DIALOG scores routinely collected from East London NHS Foundation Trust. We aimed to (a) examine changes in DIALOG scores through the patient journey (‘assessment’, ‘review’ and ‘discharge’); and (b) assess the impact of community mental health (CMH) transformation by comparing pre- and post-DIALOG scores. We analysed 11 198 DIALOG scores from 5007 patients in 2018–2019 and 2021–2022.

Results

DIALOG scores improved across treatment stages in both years. There was no clear difference pre- and post-CMH transformation, although in 2021–2022 there were lower satisfaction scores at referral.

Clinical implications

DIALOG showed sensitivity to change, supporting the utility of this scale in the evaluation of mental health services. The impact of CMH transformation was difficult to assess, due to potential confounders such as the COVID-19 pandemic. Routinely collected DIALOG data can help evaluate patient outcomes over time and inform service improvements.

Information

Type
Original Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of Royal College of Psychiatrists

In recent years there has been growing interest in the use of patient-reported outcome measures (PROMs) and patient-reported experience measures (PREMs) in healthcare settings. Such usage includes routinely collected measures for the evaluation of services such as the National Health Service (NHS) in the UK, although there are challenges concerning their effective implementation. Reference Bull, Teede, Watson and Callander1Reference Bull and Callander3 PROMs focus on targeted outcomes Reference Gelkopf, Mazor and Roe4Reference Nelson, Eftimovska, Lind, Hager, Wasson and Lindblad6 whereas PREMs focus on the experience of receiving the service. Reference Jamieson Gilmore, Corazza, Coletta and Allin7 PROMs can focus on individual conditions and/or symptoms, or on the whole person, measuring either their recovery or quality of life.

DIALOG

DIALOG is an 11-item scale that measures patient satisfaction across 11 domains. Reference Priebe and Bird8Reference Bhattacharya, Priebe and Bird11 The first eight items measure satisfaction with different life outcomes (PROMs), whereas the final three measure satisfaction about treatment/support received (PREMs). DIALOG therefore encompasses both PROM and PREM components, in a manner intended to make routine patient–clinician meetings more effective. Reference Priebe, McCabe, Bullenkamp, Hansson, Lauber and Martinez-Leal12,Reference Priebe, Golden, McCabe and Reininghaus13 The DIALOG scale forms the basis of a wider care-planning tool, called DIALOG+, which is an intervention to support patients and clinicians in co-producing a care plan focusing on each of the quality-of-life domains, with the relative importance of domains determined by the patient using the DIALOG scale. Reference Jubokowa9,10,14 For DIALOG+, in addition to rating their satisfaction for each domain using the DIALOG scale, patients are invited to discuss goals and co-produce an action plan with the clinician, based on the principles of solution-focused therapy, to enable an improvement in the domain being discussed. Reference Jubokowa9,15

East London NHS Foundation Trust (ELFT) was one of the first NHS trusts to incorporate DIALOG and DIALOG+ care planning as a part of their Care Programme Approach, soon after the development of the tool in 2017. Reference Mosler, Priebe and Bird16 The use of DIALOG (including DIALOG+) spread across a number of NHS trusts, including its acceptance as the preferred PROM for adult mental health services in London. 17 The community mental health (CMH) transformation was proposed and implemented as a part the NHS Long Term Plan (2019), 18 and was intended to improve care pathways with locally integrated multidisciplinary teams, using whole-person and whole-population health approaches. DIALOG+ co-produced care planning resonated with the CMH transformation care planning ethos. Subsequently, NHS England identified DIALOG as one of the three recommended outcome measures for CMH. 19 NHS England stated that services should work towards the routine use of DIALOG+ to support care planning, and of DIALOG for ongoing monitoring within mental health services. This dual purpose – outcome measurement and care planning – has advantages for implementation and reducing clinical burden, but potentially complicates interpretation of outcome scores.

Existing evidence

Although some studies have been conducted on DIALOG and DIALOG+, there is only limited understanding of their broader utility in the evaluation of mental health services through routinely collected data. For example, several trial-based studies have looked at DIALOG in populations with psychosis and severe mental illness. Reference Omer, Golden and Priebe20Reference Priebe, Kelley, Omer, Golden, Walsh and Khanom22 A separate study from ELFT – studying the period from January 2017 to December 2019 – evaluated routinely collected DIALOG data and found a trend of improving scores over time. Reference Mosler, Priebe and Bird16 However, this study analysed DIALOG scores over the course of treatment (using five time points) but did not distinguish the beginning and end of the patient journey (e.g. in terms of initial assessment or interim scores captured during review or at discharge). In our study, we took a different approach and looked at changes in pooled DIALOG scores through the stages of treatment (assessment, review and discharge), to capture a clear beginning and end to the patient journey at the population level.

Additionally, we aimed to compare routinely collected data pre- and post-CMH transformation, to assess their impact on DIALOG scores. This also offered us the opportunity to explore whether the pooled DIALOG scores detected any changes in the needs or satisfaction levels of the population pre-treatment (i.e. at the assessment stage before and after the COVID-19 pandemic).

Current study

We analysed anonymised, routinely collected data from RiO, ELFT’s electronic healthcare record system. We evaluated two time periods (financial years 2018–2019 and 2021–2022), capturing data pre- and post-CMH transformation. The CMH transformation was implemented in autumn 2019. These two time periods were selected to avoid data collection challenges during the height of the COVID-19 pandemic lockdown restrictions. We evaluated the data by conducting a quantitative pre–post observational analysis of the mental health service data from RiO. Apart from comparing the two time periods, the evaluation allowed us to understand whether DIALOG scores had changed over time and along the patient journey. We observed 7568 DIALOG scores in this study; this large number resulted, in part, because ELFT was an early adopter of DIALOG as both a PROM/PREM and a care-planning tool.

Our evaluation explored patient outcomes and experiences over time (in terms of both financial year and treatment stage) while controlling for several other variables (including demographic variables and protected characteristics: age, ethnicity, gender and index of multiple deprivation). This also helped us to understand the strengths and limitations of DIALOG as a routinely collected measure embedded within services.

This evaluation required close collaboration with ELFT clinicians, managers, patients, carers and data analysts to develop the analyses plan, understand data availability and quality, pathways for data input and procedures for data governance. This paper reports and discusses the findings related to treatment stage and financial year. A separate linked paper will report and discuss the results related to demographic variables and protected characteristics. This separate, linked paper will include the ways in which findings relate to the patient and carer race equality framework. 23

Method

ELFT commissioned the University of Plymouth to evaluate routinely collected community mental health team (CMHT) data, to assess the impact of the CMH transformation 24 as recommended by NHS England. As a part of the service evaluation, routinely collected DIALOG scores from CMHT services were analysed. ELFT’s business analysis team carried out a search of electronic patient records stored in RiO for DIALOG scores recorded for the identified patient group within the periods under investigation.

Design and data sources

Our evaluation used a quantitative pre–post observational design, with two cross-sectional time periods (financial years 2018–2019 and 2021–2022). The purpose of this evaluation was to assess population-level scores and changes in pooled outcome or quality-of-life measures and experience, rather than changes in individual patients. As part of the CMH transformation there was a greater focus on integration with primary care, and the post-transformation community teams had a broader scope, including what was previously the remit of primary care liaison teams. DIALOG scores from CMHTs and primary care liaison teams for 2018–2019, and from the ‘transformed’ community or neighbourhood mental health team in 2021–2022 (which offered the function of both the previous teams), were considered in scope. The patient group were all adults aged 18 years and above by the start of the first financial year (2018–2019) and with at least one DIALOG score in one of the two study periods (regardless of when they were first assessed or finally discharged).

The data were collected from three London boroughs serviced by ELFT: City and Hackney, Tower Hamlets and Newham. We analysed 11 198 DIALOG scores, 5294 for 2018–2019 and 5904 for 2021–2022. The number of unique patients (i.e. individual patients) in the study was 5007; 2693 unique patients were analysed in 2018–2019 and 3161 in 2021–2022 (because some patients were present in both data periods, the total number of unique patients is lower than the sum of unique patients in each financial year).

Of note, not all domains were filled in at each submission. We defined ‘stage of treatment’ as ‘assessment’ for new referrals, ‘review’ for ongoing treatment and as ‘discharge’ for end of treatment. We pooled analysis by stage of treatment at the DIALOG domain level for some analyses.

Measures

We compared pooled, pseudo-anonymised DIALOG scores along with protected characteristics and other demographic variables (age, ethnicity, gender and index of multiple deprivation (IMD) decile). We analysed data within and across the two time periods described above, and by stage of treatment (assessment, review and discharge).

Materials

The DIALOG scale is presented in Table 1. Each DIALOG domain item is scored using a Likert scale ranging from 1 (totally dissatisfied) to 7 (totally satisfied).

Table 1 DIALOG scale. The items are scored on a scale of 1–7, where 1 is totally dissatisfied and 7 is totally satisfied

Analyses

Two types of analyses were carried out using the R statistical package (2024 version) for Windows and Ubuntu (R Core Team, The R Foundation, Vienna, Austria; https://www.r-project.org/). 25 Because the analytical framework Reference Bhattacharya, Priebe and Bird11 does not specify any statistical analyses of pooled DIALOG data, we developed the following approach as part of our evaluation.

  1. (a) We performed descriptive statistical analyses of pooled DIALOG scores, including means and 95% confidence intervals, across a range of variables for each DIALOG domain, financial year and stage of treatment. We report means and CIs for the two financial years and three stages of treatment.

  2. (b) We used multiple logistic regressions on the DIALOG domains, where we converted the raw scores into a binary variable of ‘satisfied’ (scores of 4–7) and ‘dissatisfied’ (1–3). Thereafter, for each DIALOG domain we estimated a multivariable logistic regression, with the odds of reporting a ‘satisfied’ DIALOG domain score as the outcome variable and the following explanatory variables: stage of treatment, financial year, age, ethnicity, gender and IMD decile (we cross-referred DIALOG scores from individuals with the IMD scores generated from postcode information, using openly available government data 26 ). The interaction between treatment stage and financial year was also analysed. If models including the interaction term performed significantly better at explaining the data than those without, the former were selected and reported – otherwise the latter were selected and reported. The model selection was conducted using analysis of deviance significance tests.

There was clinical importance in understanding whether patients had moved from being dissatisfied to satisfied (or vice versa) over the course of the patient journey from assessment to discharge. Logistic regression measures the odds of these satisfaction changes at the population level. This provides a more clinically useful measure than more abstract changes in level of satisfaction along an ordinal scale, while also being a more statistically strict form of analysis. However, the original raw DIALOG scores are also reported in the descriptive statistics (means and 95% confidence intervals), to provide a full report of the data-set. The 95% confidence intervals provide an indication of where mean scores are significantly different. Additional significance testing on the descriptive statistics was deemed redundant, because the logistic regressions provide a more statistically sophisticated and robust analysis of potential effects.

Results

We analysed routinely collected data from 5007 patients with at least one DIALOG score and reported at least in one domain: a total of 11 198 DIALOG assessments across both years. This was then split, because 2693 patients had a total of 5294 assessments in financial year 2018–2019 and 3161 had a total of 5904 in financial year 2021–2022. The DIALOG scores were collected routinely for patients receiving adult CMH treatment from the CMHT and primary care liaison teams in the east London boroughs of City and Hackney, Tower Hamlets and Newham in 2018–2019, and the ‘transformed’ CMHTs in the same boroughs in 2021–2022, with the care being delivered by ELFT. Routinely collected DIALOG scores linked to treatment stage were obtained from electronic patient records, anonymised and pooled for the analysis.

First, we pooled both time periods to analyse differences in scores across stages of treatment. Figure 1 shows that DIALOG scores across the board tended to improve with duration within the service; this was true both pre- and post- pandemic, and is statistically significant (according to 95% confidence intervals).

Fig. 1 Mean DIALOG scores for each domain, split by treatment stage across 2018–2019 and 2021–2022 combined. Error bars are 95% confidence intervals; bars with non-overlapping confidence intervals can be interpreted as significantly different. There were n = 4193 sets of DIALOG assessment scores, n = 6764 sets of DIALOG review scores and n = 240 sets of DIALOG discharge scores. MH, mental health; PH, physical health; JS, job situation; AC, accommodation; LA, leisure activities; RS, relationship with partner/family; FS, friendships; PS, personal safety; MD, medication; PR, the practical help you receive; MP, meetings with mental health professionals.

These results show that patient satisfaction improved over the categorical stages of the patient journey, from assessment (time of referral), through reviews (mid-treatment) and, finally, to discharge (end of treatment). Because collection of DIALOG scores at discharge was carried out less routinely, the number of discharge scores (240) was lower than that at both assessment (4193) and review (6764). This may also indicate that most patients remained within the services during the evaluation period.

Figure 2 shows analyses of these same scores by year, allowing for comparison of pre- and post-CMH transformation. As in Fig. 1, Fig. 2 shows a trend of increasing satisfaction by treatment stage, but also some marginal evidence of a decrease in satisfaction from pre- to post-transformation. We identified an increase in mental health need in those referred to ELFT services from the community in 2021–2022 compared with 2018–2019, as evident from poorer satisfaction with mental health at the time of referral. There was also an apparent reduction in satisfaction for physical health at assessment from 2018–2019 to 2021–2022. However, these differences were not significant at the 95% confidence interval level. The small number of observations at discharge increases uncertainty in the estimates, such that differences in mean scores at discharge were not statistically significant between the two time periods. Additionally, these results suggest an increase in satisfaction post-transformation for people in contact (reviews) with ELFT mental health services (for mental health, physical health, leisure activities, friendships, personal safety and medication).

Fig. 2 DIALOG scores by stage (assessment, review and discharge) and year (2018 refers to 2018–2019 and 2021 refers to 2021–2022). Error bars are 95% confidence intervals; bars with non-overlapping confidence intervals can be interpreted as significantly different. In this subgroup, there were n = 1515 and 2678 sets of DIALOG assessment scores in 2018 and 2021, respectively, n = 3643 and 3121 sets of DIALOG review scores in 2018 and 2021, respectively and n = 136 and 104 sets of DIALOG discharge scores in 2018 and 2021, respectively. See Fig. 1 for list of abbreviations used.

Figures 1 and 2 present descriptive univariate analyses, and do not address potential confounding variables. Below, we present results from multivariable regressions that estimate DIALOG domain scores, controlling for age, ethnicity, gender and index of multiple deprivation. For ease of interpretation, the regression results are split across three separate figures (see Supplementary Fig. 1 for a combined plot). Figure 3 reports the regression results for discharge compared with assessment, comparing 2021–2022 scores against 2018–2019.

Fig. 3 Results of multiple regressions on DIALOG scores by year (2018 refers to 2018–2019 and 2021 refers to 2021–2022) and treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals; if whiskers are >1, this variable is significantly associated with greater odds of being satisfied; and if <1, associated with lower odds of being satisfied; intermediate values are not significant. Predictor variables are all labelled to show what they are being compared against, e.g. ‘2021 v. 2018’ means that 2021–2022 is the predictor variable level and 2018–2019 is what we are comparing it against. See Fig. 1 for list of abbreviations used.

Results controlling for demographic characteristics are largely consistent with the findings from descriptive univariate analyses by treatment stage. The odds of patients being satisfied were higher at discharge than assessment, for all DIALOG domains apart from personal safety. However, for most variables there was no difference in satisfaction between the two time periods, although we did observe a statistically significant increase in satisfaction with friendships and personal safety between 2018–2019 and 2021–2022.

Figure 4 shows that the odds of satisfaction were higher at review (midway through treatment) compared with those at assessment (initiation of treatment) for all 11 DIALOG domains.

Fig. 4 Results of multiple regressions on DIALOG scores by treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals; if whiskers are >1, this variable is significantly associated with greater odds of being satisfied; and if <1, this is associated with lower odds of being satisfied; intermediate values are not significant. See Fig. 1 for list of abbreviations used.

We also estimated models, including an interaction term for year × treatment stage, to test whether differences across treatment stage exhibited different patterns across the different years. There were only three DIALOG domains where the regression model that included an interaction term, between treatment stage and year, performed significantly better than the regression model with no interaction term (i.e. explains significantly more of the data). These are reported in Fig. 5. The results show a significant negative interaction between review (versus assessment) and year (2021–2022 versus 2018–2019) for friendships, practical help received and meetings with mental health professionals. This means that, in 2021–2022 there was a reduced increase in satisfaction, from assessment to review, for these domains (patient satisfaction, however, still improved from assessment to review – see Fig. 4).

Fig. 5 Results of multiple regressions on DIALOG scores by year and treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals. The interactions show how changes in satisfaction by treatment stage differ between the two years. The predictor variables in the legend are labelled to show what they are being compared against (assessment for review and discharge: 2018–2019 for 2021–2022). Again, the plots are interpreted as greater odds for whiskers >1, and lower odds for those <1. See Fig. 1 for list of abbreviations used.

Discussion

Our results are consistent with earlier findings by Mosler et al, Reference Mosler, Priebe and Bird16 and show that routinely collected DIALOG data can provide a sensitive and useful tool for detecting changes in patient satisfaction associated with contact with mental health services. Our results do not merely replicate the Mosler et al study – they show that DIALOG patient satisfaction measures improved with treatment stage. In other words, the treatment journey through mental health services was associated with improvements to patient satisfaction for both outcomes and services. Moreover, patients were more satisfied across the various quality-of-life domains, not just mental health, once they were receiving support (i.e. already assessed and had had at least one review, compared with initial contact of assessment). There was further improvement in satisfaction at the point of discharge (compared with review stage). The only exception to this was for personal safety at the point of discharge in 2021–2022, in which the observed improvement was not statistically significant. We observed improvement across the eight PROM domains as well as the three PREM domains. Improvement was observed particularly in the practical help domain throughout the stages of treatment (PREM scores were not reviewed in the study by Mosler et al).

Satisfaction across the quality-of-life domains, as measured by DIALOG, were not markedly different between the two years evaluated. However, there were some subtle exceptions to this. Post-pandemic, there was a perceived reduction in satisfaction with mental health at the point of initial assessment; this can be understood potentially as an increase in mental health needs in the community during the pandemic. In the multiple regressions, after controlling for other variables, friendships and practical help received were the only two DIALOG domains in which there was an improvement in patient satisfaction between 2018–2019 and 2021–2022, whereas there was no evidence of any change for the other nine domains. We know that the pandemic and lockdowns altered social interactions, and that many community centres and resources were shut down.

It is difficult to draw clear inferences about the impact of the CMH transformation from these results, partly as a result of the COVID-19 pandemic producing a confound. The pandemic is known to have had a detrimental impact on mental health and well-being, Reference Pierce, Hope, Ford, Hatch, Hotopf and John27Reference Chen and Wang30 while also limiting face-to-face contact in both healthcare settings and everyday social interactions Reference Anderson, Walsh, Anderson and Burnley31Reference Schneiders, Mackworth-Young and Cheah34 (e.g. greater use of online support, less face-to-face contact, less community interaction). It is also due to the well-recognised limitations of uncontrolled whole-service evaluations detailed below. Measuring the impact of service reorganisation is challenging, and attempts to capture the impact of reconfiguration have often not been successful. Reference Giacco, Bird, Ahmad, Bauer, Lasalvia and Lorant35

We have identified several limitations in our evaluation. First, DIALOG response rates were low: scores were estimated to be available for less than 20% of ELFT CMHT patients for both 2018–2019 and 2021–2022. We cannot rule out the possibility of systematic differences in ELFT patients completing DIALOG questionnaires compared with those who did not. We also observed a comparatively low n figure at discharge compared with assessment and review, indicating attrition over time. From a broader perspective, working with routinely collected natural data is more prone to confounding effects than with formal research data, with less scope for incorporating appropriate controls. Reference Nørgaard, Ehrenstein and Vandenbroucke36,Reference Sauer, Chen, Hyland, Girbes, Elbers and Celi37 However, these limitations need to be balanced against the benefits of leveraging real-life data to evaluate and improve service provision, including larger sample sizes, better generalisability and lower costs. Reference Bull, Teede, Watson and Callander1,Reference Sauer, Chen, Hyland, Girbes, Elbers and Celi37,Reference Von Gerich, Peltonen, Séroussi, Weber, Dhombres, Grouin, Liebe and Pelayo38

We were also limited by the nature and intended purpose of DIALOG data, which are primarily intended to aid in care planning during active mental health treatment, and to judge the success of treatment in meeting an individual’s goals for care. These data are informative, but challenging to use, for causal inference or evaluation purposes at an aggregate level, because they have no standard study period, study population or data collection schedule. This makes it difficult to establish baselines and controls for studying differences in differences. Further longitudinal analysis of DIALOG is needed to better understand its utility as a tool for evaluation. Nevertheless, as our findings demonstrate, routinely collected mental health data can provide a useful tool for understanding the needs of the population being referred, changes through the patient journey and other aspects of population evaluation.

Although we did find that patients’ quality of life had improved overall, especially for those who continued to engage with treatment, our study does not provide any answers as to ‘why’. Our study looked at the overall change in PROM and PREM through the patient journey while receiving care from CMH services. Quality of life is affected by a multitude of variables both inside and outside the health system, and contact with services potentially forms only a fraction of this experience. The improvement in quality-of-life domains extended beyond mental health. There can be a range of hypotheses as to how and why this happened, including improvement in the mental health domain having a positive impact on other quality-of-life domains, as well as the impact of care not being restricted to one’s mental health. However, we cannot infer a direct causality between the care received and improvement in DIALOG scores.

We propose the need for greater focus on routine outcome and experience data-gathering, with real-time analysis of the data shared at an individual patient, team or service level, as well as across the organisation and between healthcare providers. A significant focus of the CMH transformation work was around improving access and care coordination, which DIALOG is less able to capture. Access and other parameters need to be considered in conjunction with the pooled DIALOG scores in assessing overall quality of impact across the population. In other words, we are unable to gauge how many people might be missing out on ELFT mental health services, and what challenges they may face. Future research could investigate such potential underserved populations, with a particular focus on health inequalities. Another suggested avenue for future research is to compare DIALOG scores with other routine measures of improvement (e.g. Health of the Nation Outcome Scales), to check whether they align. This has already been carried out in a small-scale study at service level in the same organisation. Reference Butt, Walls and Bhattacharya39 Finally, our study is not a formal evaluation of the measurement properties of DIALOG, although it would be useful to investigate properties such as construct validity, responsiveness and reliability. Reference Mokkink, Prinsen, Bouter, Vet and Terwee40

In conclusion, we conducted a quantitative evaluation of patient outcomes and experiences within ELFT mental health services during the CMH transformation, using routinely collected DIALOG scores. Our analyses investigated the change in DIALOG scores by treatment stage for two different years. Our results showed that pooled, routinely collected DIALOG data can provide a useful measure of changes in patient satisfaction across the 11 domains (both PROM and PREM). The changes or improvements in DIALOG scores were similar across the two study years, and scores improved over the course of the patient journey from assessment to discharge in both time periods. Although this effect is not causal, it is suggestive of a positive impact of treatment on quality-of-life domains, including patient experience. Further research may consider the use of DIALOG as an assessment tool in a controlled study; comparisons of quality of life among people in mental health treatment, especially specific interventions, compared with those who are not; or variations in baseline scores or responsiveness to treatment in certain subgroups, including historically underserved populations. Overall, our results highlight both the strengths and limitations of routinely collected DIALOG data, as well as PROM/PREM data more broadly.

About the authors

S.G.S. is a senior research fellow in applied healthcare at the University of Plymouth Community & Primary Care Research Centre, Plymouth, UK and PenARC, Plymouth, UK. R.Bh. is a consultant psychiatrist and Clinical Lead for Mental Health Payment and Outcomes for East London NHS Foundation Trust, London, UK and Honorary Associate Clinical Professor at Warwick Medical School, University of Warwick, Coventry, UK. K.S. is a data analyst in healthcare and was a researcher at the University of Plymouth, Plymouth, UK at the time of this study. A.S. is a consultant psychiatrist at North East London NHS Foundation Trust, London, UK. P.S. was a researcher in psychology and applied healthcare at the University of Plymouth, Plymouth, UK at the time of this study. R.By. is Professor in Primary Care Research at the University of Plymouth, Plymouth, UK, Head of the University of Plymouth Community & Primary Care Research Centre, Plymouth, UK and Deputy Director, PenARC, Plymouth, UK.

Supplementary material

The supplementary material is available online at https://doi.org/10.1192/bjb.2026.10215.

Data availability

The data underlying this study are derived from anonymised patient records from East London NHS Foundation Trust. Due to the sensitive nature of NHS mental healthcare data and information governance restrictions, these data are not publicly available. Access to the data is subject to appropriate approvals from the Trust and relevant governance bodies, and may be considered on reasonable request, subject to data sharing agreements and ethical approval.

Acknowledgements

Prof. Stefan Priebe for his advice on the paper; Prof. Frank Rohricht, Medical Director for Research and Innovation from ELFT, who commissioned the initial analysis and agreed to further analysis; and Thomas Nicholas, Associate Director for Business Intelligence and Analytics, ELFT, for help with data-capture.

Author contributions

S.G.S., K.S., R.By. and R.Bh. conceived the study, including its rationale, aims and methodology. K.S., P.S. and S.G.S. conducted the initial analyses. S.G.S. completed the analyses with input from R.Bh. and A.S. S.G.S. drafted the first version of the manuscript, and S.G.S., R.Bh. and A.S. worked on the revised version. All authors had input into and approved the final version.

Funding

East London NHS Foundation Trust commissioned the University of Plymouth to conduct this evaluation. Funding was awarded to the University of Plymouth following a competitive tender process. S.G.S., R.By., K.S. and P.S. were additionally funded and supported by the National Institute for Health & Care Research Applied Research Collaboration South West Peninsula. R.By. and A.S. contributed to the evaluation as University of Plymouth partners; they are employed by East London Foundation Trust.

Declaration of interest

R.Bh. is a member of the BJPsych Bulletin editorial board; he did not take part in the review or decision-making process of this paper.

Ethical standards

East London NHS Foundation Trust commissioned the University of Plymouth to evaluate routinely collected CMHT data, to assess the impact of the CMH transformation, as recommended by NHS England. Because these routinely collected data were part of a commissioned local evaluation rather than research (i.e. classed as service evaluation rather than research under the UK Policy Framework for Health and Social Care Research), ethical approval was not required (as per Health Reimbursement Arrangement standards, and was agreed by the ELFT Ethics Committee). The data-sharing agreement for the evaluation was detailed within the terms of the contract.

References

Bull, C, Teede, H, Watson, D, Callander, EJ. Selecting and implementing patient-reported outcome and experience measures to assess health system performance. JAMA Health Forum 2022; 3: e220326.10.1001/jamahealthforum.2022.0326CrossRefGoogle ScholarPubMed
Benson, T. Why it is hard to use PROMs and PREMs in routine health and care. BMJ Open Qual 2023; 12: e002516.10.1136/bmjoq-2023-002516CrossRefGoogle ScholarPubMed
Bull, C, Callander, EJ. Current PROM and PREM use in health system performance measurement: still a way to go. Patient Exp J 2022; 9: 1218.10.35680/2372-0247.1664CrossRefGoogle Scholar
Gelkopf, M, Mazor, Y, Roe, D. A systematic review of patient-reported outcome measurement (PROM) and provider assessment in mental health: goals, implementation, setting, measurement characteristics and barriers. Int J Qual Health Care 2022; 34: ii1327.10.1093/intqhc/mzz133CrossRefGoogle Scholar
Kendrick, T, El-Gohary, M, Stuart, B, Gilbody, S, Churchill, R, Aiken, L, et al. Routine use of patient reported outcome measures (PROMs) for improving treatment of common mental health disorders in adults. Cochrane Database Syst Rev 2016: 7: CD011119.Google ScholarPubMed
Nelson, EC, Eftimovska, E, Lind, C, Hager, A, Wasson, JH, Lindblad, S. Patient reported outcome measures in practice. BMJ 2015; 350: g7818.10.1136/bmj.g7818CrossRefGoogle ScholarPubMed
Jamieson Gilmore, K, Corazza, I, Coletta, L, Allin, S. The uses of Patient Reported Experience Measures in health systems: a systematic narrative review. Health Policy 2023; 128: 110.10.1016/j.healthpol.2022.07.008CrossRefGoogle ScholarPubMed
Priebe, S, Bird, V. DIALOG Scale – Analytical Framework for Mental Health Services. East London NHS Foundation Trust, 2019 (https://www.elft.nhs.uk/sites/default/files/DIALOG%20Analytical%20Framework.pdf [accessed 13 Mar 2024]).Google Scholar
Jubokowa, B. A Guide to Complete Dialog+ Assessment Form on RiO. Transformation Partners in Health and Care, 2019 (https://www.transformationpartners.nhs.uk/wp-content/uploads/2019/10/Technical-User-Guide-to-complete-Dialog_Draft.docx [accessed 25 Mar 2025]).Google Scholar
Transformation Partners in Health and Care. Outcome Measures (DIALOG and DIALOG+). Transformation Partners in Health and Care, n.d. (https://www.transformationpartners.nhs.uk/programmes/mental-health-transformation/support-for-adults/new-models-of-community/outcome-measures-dialog-and-honos/ [accessed 25 Mar 2025]).Google Scholar
Bhattacharya, R, Priebe, S, Bird, V. DIALOG Scale – Analytical Framework for Mental Health Services. East London NHS Foundation Trust, n.d.Google Scholar
Priebe, S, McCabe, R, Bullenkamp, J, Hansson, L, Lauber, C, Martinez-Leal, R, et al. Structured patient–clinician communication and 1-year outcome in community mental healthcare: cluster randomised controlled trial. Br J Psychiatry 2007; 191: 420–6.10.1192/bjp.bp.107.036939CrossRefGoogle ScholarPubMed
Priebe, S, Golden, E, McCabe, R, Reininghaus, U. Patient-reported outcome data generated in a clinical intervention in community mental health care - psychometric properties. BMC Psychiatry 2012; 12: 113.10.1186/1471-244X-12-113CrossRefGoogle Scholar
East London NHS Foundation Trust. DIALOG+. ELNHSFT, 2023 (https://www.elft.nhs.uk/dialog [accessed 10 Mar 2025]).Google Scholar
Healthy London Partnership. London Mental Health Transformation Programme – DIALOG Operational Manual. Healthy London Partnership, n.d.Google Scholar
Mosler, F, Priebe, S, Bird, V. Routine measurement of satisfaction with life and treatment aspects in mental health patients – the DIALOG scale in East London. BMC Health Serv Res 2020; 20: 1020.10.1186/s12913-020-05840-zCrossRefGoogle Scholar
Central and North West London NHS Foundation Trust. Launching DIALOG+ in Older Adult Community Mental Health Teams. CNWLNHSFT, 2023 (https://www.cnwl.nhs.uk/news/launching-dialog-older-adult-community-mental-health-teams [accessed 10 Mar 2025]).Google Scholar
National Health Service. Fit for the Future. NHS, 2020 (https://www.longtermplan.nhs.uk [accessed 1 Feb 2023]).Google Scholar
NHS England. Implementation Guidance 2024 – Psychological Therapies for Severe Mental Health Problems. NHS England, 2024 (https://www.england.nhs.uk/long-read/implementation-guidance-2024-psychological-therapies-for-severe-mental-health-problems/ [accessed 5 Mar 2025]).Google Scholar
Omer, S, Golden, E, Priebe, S. Exploring the mechanisms of a patient-centred assessment with a solution focused approach (DIALOG+) in the community treatment of patients with psychosis: a process evaluation within a cluster-randomised controlled trial. PLOS One 2016; 11: e0148415.10.1371/journal.pone.0148415CrossRefGoogle Scholar
Slatina Murga, S, Janković, S, Muhić, M, Sikira, H, Burn, E, Priebe, S, et al. Effectiveness of a structured intervention to make routine clinical meetings therapeutically effective (DIALOG+) for patients with depressive and anxiety disorders in Bosnia and Herzegovina: a cluster randomised controlled trial. Psychiatry Res Commun 2021; 1: 100010.10.1016/j.psycom.2021.100010CrossRefGoogle ScholarPubMed
Priebe, S, Kelley, L, Omer, S, Golden, E, Walsh, S, Khanom, H, et al. The effectiveness of a patient-centred assessment with a solution-focused approach (DIALOG+) for patients with psychosis: a pragmatic cluster-randomised controlled trial in community care. Psychother Psychosomat 2015; 84: 304–13.10.1159/000430991CrossRefGoogle ScholarPubMed
NHS England. Patient and Carer Race Equality Framework. NHS England, 2023 (https://www.england.nhs.uk/mental-health/advancing-mental-health-equalities/pcref/ [accessed 7 Jun 2024]).Google Scholar
National Health Service. NHS Mental Health Implementation Plan 2019/20–2023/24. NHS, 2019.Google Scholar
R Core Team. R: A Language and Environment for Statistical Computing. R Core Team, 2024.Google Scholar
Ministry of Housing, Communities & Local Government. English Indices of Deprivation 2019. Ministry of Housing, Communities & Local Government, 2019 (https://imd-by-postcode.opendatacommunities.org/imd/2019 [accessed 10 Mar 2025]).Google Scholar
Pierce, M, Hope, H, Ford, T, Hatch, S, Hotopf, M, John, A, et al. Mental health before and during the COVID-19 pandemic: a longitudinal probability sample survey of the UK population. Lancet Psychiatry 2020; 7: 883–92.10.1016/S2215-0366(20)30308-4CrossRefGoogle ScholarPubMed
O’Connor, RC, Wetherall, K, Cleare, S, McClelland, H, Melson, AJ, Niedzwiedz, CL, et al. Mental health and well-being during the COVID-19 pandemic: longitudinal analyses of adults in the UK COVID-19 Mental Health & Wellbeing study. Br J Psychiatry 2021; 218: 326–33.10.1192/bjp.2020.212CrossRefGoogle ScholarPubMed
Close, J, Spicer, SG, Nicklin, LL, Lloyd, J, Whalley, B, Lloyd, H. Gambling and gaming in the United Kingdom during the COVID-19 lockdown. COVID 2022; 2: 87101.10.3390/covid2020007CrossRefGoogle Scholar
Chen, DT-H, Wang, Y-J. Inequality-related health and social factors and their impact on well-being during the COVID-19 pandemic: findings from a national survey in the UK. Int J Environ Res Public Health 2021; 18: 1014.10.3390/ijerph18031014CrossRefGoogle ScholarPubMed
Anderson, J, Walsh, J, Anderson, M, Burnley, R. Patient satisfaction with remote consultations in a primary care setting. Cureus 2021: 13: e17814.Google Scholar
Lifford, KJ, Grozeva, D, Cannings-John, R, Quinn-Scoggins, H, Moriarty, Y, Gjini, A, et al. Satisfaction with remote consultations in primary care during COVID-19: a population survey of UK adults. Br J Gen Pract 2023: 74: e96103.10.3399/BJGP.2023.0092CrossRefGoogle Scholar
Turner, A, Scott, A, Horwood, J, Salisbury, C, Denholm, R, Scott, L, et al. Maintaining face-to-face contact during the COVID-19 pandemic: a longitudinal qualitative investigation in UK primary care. BJGP Open 2021; 5: BJGPO.2021.0036.10.3399/BJGPO.2021.0036CrossRefGoogle ScholarPubMed
Schneiders, ML, Mackworth-Young, CRS, Cheah, PY. Between division and connection: a qualitative study of the impact of COVID-19 restrictions on social relationships in the United Kingdom. Wellcome Open Res 2022; 7: 6.10.12688/wellcomeopenres.17452.1CrossRefGoogle ScholarPubMed
Giacco, D, Bird, VJ, Ahmad, T, Bauer, M, Lasalvia, A, Lorant, V, et al. The same or different psychiatrists for in- and out-patient treatment? A multi-country natural experiment. Epidemiol Psychiatr Sci 2020; 29: e10.10.1017/S2045796018000732CrossRefGoogle Scholar
Nørgaard, M, Ehrenstein, V, Vandenbroucke, JP. Confounding in observational studies based on large health care databases: problems and potential solutions – a primer for the clinician. Clin Epidemiol 2017; 9: 185–93.10.2147/CLEP.S129879CrossRefGoogle ScholarPubMed
Sauer, CM, Chen, L-C, Hyland, SL, Girbes, A, Elbers, P, Celi, LA. Leveraging electronic health records for data science: common pitfalls and how to avoid them. Lancet Digit Health 2022; 4: e8938.10.1016/S2589-7500(22)00154-6CrossRefGoogle Scholar
Von Gerich, H, Peltonen, L-M. Assessment of health service quality through electronic health record – a scoping review. In Studies in Health Technology and Informatics (eds Séroussi, B, Weber, P, Dhombres, F, Grouin, C, Liebe, JD, Pelayo, S, et al). IOS Press, 2022.Google Scholar
Butt, MF, Walls, D, Bhattacharya, R. Do patients get better? A review of outcomes from a crisis house and home treatment team partnership. BJPsych Bull 2019; 43: 106–11.10.1192/bjb.2018.105CrossRefGoogle Scholar
Mokkink, LB, Prinsen, CAC, Bouter, LM, Vet, HCWD, Terwee, CB. The COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) and how to select an outcome measurement instrument. Braz J Phys Ther 2016; 20: 105–13.10.1590/bjpt-rbf.2014.0143CrossRefGoogle ScholarPubMed
Figure 0

Table 1 DIALOG scale. The items are scored on a scale of 1–7, where 1 is totally dissatisfied and 7 is totally satisfied

Figure 1

Fig. 1 Mean DIALOG scores for each domain, split by treatment stage across 2018–2019 and 2021–2022 combined. Error bars are 95% confidence intervals; bars with non-overlapping confidence intervals can be interpreted as significantly different. There were n = 4193 sets of DIALOG assessment scores, n = 6764 sets of DIALOG review scores and n = 240 sets of DIALOG discharge scores. MH, mental health; PH, physical health; JS, job situation; AC, accommodation; LA, leisure activities; RS, relationship with partner/family; FS, friendships; PS, personal safety; MD, medication; PR, the practical help you receive; MP, meetings with mental health professionals.

Figure 2

Fig. 2 DIALOG scores by stage (assessment, review and discharge) and year (2018 refers to 2018–2019 and 2021 refers to 2021–2022). Error bars are 95% confidence intervals; bars with non-overlapping confidence intervals can be interpreted as significantly different. In this subgroup, there were n = 1515 and 2678 sets of DIALOG assessment scores in 2018 and 2021, respectively, n = 3643 and 3121 sets of DIALOG review scores in 2018 and 2021, respectively and n = 136 and 104 sets of DIALOG discharge scores in 2018 and 2021, respectively. See Fig. 1 for list of abbreviations used.

Figure 3

Fig. 3 Results of multiple regressions on DIALOG scores by year (2018 refers to 2018–2019 and 2021 refers to 2021–2022) and treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals; if whiskers are >1, this variable is significantly associated with greater odds of being satisfied; and if <1, associated with lower odds of being satisfied; intermediate values are not significant. Predictor variables are all labelled to show what they are being compared against, e.g. ‘2021 v. 2018’ means that 2021–2022 is the predictor variable level and 2018–2019 is what we are comparing it against. See Fig. 1 for list of abbreviations used.

Figure 4

Fig. 4 Results of multiple regressions on DIALOG scores by treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals; if whiskers are >1, this variable is significantly associated with greater odds of being satisfied; and if <1, this is associated with lower odds of being satisfied; intermediate values are not significant. See Fig. 1 for list of abbreviations used.

Figure 5

Fig. 5 Results of multiple regressions on DIALOG scores by year and treatment stage. Higher values indicate higher odds of satisfaction on each DIALOG domain. Whiskers are 95% confidence intervals. The interactions show how changes in satisfaction by treatment stage differ between the two years. The predictor variables in the legend are labelled to show what they are being compared against (assessment for review and discharge: 2018–2019 for 2021–2022). Again, the plots are interpreted as greater odds for whiskers >1, and lower odds for those <1. See Fig. 1 for list of abbreviations used.

Supplementary material: File

Spicer et al. supplementary material

Spicer et al. supplementary material
Download Spicer et al. supplementary material(File)
File 75 KB
Submit a response

eLetters

No eLetters have been published for this article.