Hostname: page-component-78c5997874-94fs2 Total loading time: 0 Render date: 2024-11-10T21:01:10.654Z Has data issue: false hasContentIssue false

Scaling up mental health care and psychosocial support in low-resource settings: a roadmap to impact

Published online by Cambridge University Press:  26 November 2020

Mark J. D. Jordans*
Affiliation:
Research & Development, War Child Holland, Amsterdam, the Netherlands Amsterdam Institute of Social Science Research, University of Amsterdam, Amsterdam, the Netherlands
Brandon A. Kohrt
Affiliation:
Division of Global Mental Health, Department of Psychiatry & Behavioral Sciences, The George Washington University
*
Author for correspondence: Mark Jordans, E-mail: mark.jordans@kcl.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

Aims

Despite recent global attention to mental health and psychosocial support services and a growing body of evidence-support interventions, few mental health services have been established at a regional or national scale in low- and middle-income countries (LMIC). There are myriad challenges and barriers ranging from testing interventions that do not target priority needs of populations or policymakers to interventions that cannot achieve adequate coverage to decrease the treatment gap in LMIC.

Method

We propose a ‘roadmap to impact’ process that guides planning for interventions to move from the research space to the implementation space.

Results

We establish four criteria and nine associated indicators that can be evaluated in low-resource settings to foster the greatest likelihood of successfully scaling mental health and psychosocial interventions. The criteria are relevance (indicators: population need, cultural and contextual fit), effectiveness (change in mental health outcome, change in hypothesised mechanism of action), quality (adherence, competence, attendance) and feasibility (coverage, cost). In the research space, relevance and effectiveness need to be established before moving into the implementation space. In the implementation space, ongoing monitoring of quality and feasibility is required to achieve and maintain a positive public health impact. Ultimately, a database or repository needs to be developed with these criteria and indicators to help researchers establish and monitor minimum benchmarks for the indicators, and for policymakers and practitioners to be able to select what interventions will be most likely to succeed in their settings.

Conclusion

A practicable roadmap with a sequence of measurable indicators is an important step to delivering interventions at scale and reducing the mental health treatment gap around the world.

Type
Special Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2020. Published by Cambridge University Press

Introduction

Mental health care in low- and middle-income countries (LMIC) is receiving increasing attention in research, practice and policy. Recent years have seen several responses to the large unmet mental health needs due to unavailable human and financial resources in LMIC. The Sustainable Development Goals (SDG) include a clear reference to mental health and the World Health Organization (WHO) is enacting a Comprehensive Mental Health Action Program. With the 2018 first-ever inter-ministerial summit on global mental health in London and a second in 2019 in the Netherlands focusing on mental health in humanitarian settings, the need for mental health care in LMIC has made its way to policy makers. Researchers have evaluated models of mental health services delivered by non-professionals, and synthesised current evidence in a series of recent reviews (van Ginneken et al., Reference van Ginneken, Tharyan, Lewin, Rao, Meera, Pian, Chandrashekar and Patel2013; Singla et al., Reference Singla, Kohrt, Murray, Anand, Chorpita and Patel2017; Kohrt et al., Reference Kohrt, Asher, Bhardwaj, Fazel, Jordans, Mutamba, Nadkarni, Pedersen, Singla and Patel2018; Purgato et al., Reference Purgato, Gross, Betancourt, Bolton, Bonetto, Gastaldon, Gordon, O'Callaghan, Papola, Peltonen, Punamaki, Richards, Staples, Unterhitzenberger, van Ommeren, de Jong, Jordans, Tol and Barbui2018). However, with all these positive developments, few mental health interventions and programmes have been brought to scale.

A major challenge is knowing what works where, for whom and how. The questions inform what interventions should be scaled and how to assure that they are effective at scale. There is variability in outcomes among and within LMIC settings even when the same intervention is used (Fuhr et al., Reference Fuhr, Weobong, Lazarus, Vanobberghen, Weiss, Singla, Tabana, Afonso, De Sa, D'Souza, Joshi, Korgaonkar, Krishna, Price, Rahman and Patel2019; Sikander et al., Reference Sikander, Ahmad, Atif, Zaidi, Vanobberghen, Weiss, Nisar, Tabana, Ain, Bibi, Bibi, Liaqat, Sharif, Zulfiqar, Fuhr, Price, Patel and Rahman2019; Dorsey et al., Reference Dorsey, Lucid, Martin, King, O’Donnell, Murray, Wasonga, Itemba, Cohen, Manongi and Whetten2020). Yet, it is not feasible to perform extensive adaptations and run new randomised controlled trials (RCT) in every single setting where a new mental health or psychosocial intervention is going to be deployed. That said, the variability in outcomes means that one cannot assume that interventions will be effective – and what may be going wrong when they are not effective – when scaled beyond the original effectiveness RCT (Kohrt et al., Reference Kohrt, El Chammay and Dossen2020). Unfortunately, there is not a minimum set of guidelines for what to monitor in the scale-up process to measure if and why or why not an intervention is working.

In a time that mental health services are increasingly being incorporated in policy and planning (Patel et al., Reference Patel, Saxena, Lund, Thornicroft, Baingana, Bolton, Chisholm, Collins, Cooper, Eaton, Hermann, Herzallah, Huang, Jordans, Kleinman, Medina Mora, Morgan, Niaz, Omigbodun, Prince, Rahman, Saraceno, Sarkar, De Silva, Singh, Stein, Sunkel and Unutzer2018), we think a roadmap is needed to provide guidance for governments and international agencies that are implementing such services both for new and existing services. Such a roadmap would outline a practicable trajectory for testing and implementing mental health and psychosocial support interventions, using a minimum set of criteria that maximise scalability and impact combined with tools to operationalise and benchmark the criteria. To meet this need, we propose a ‘roadmap to impact’ for scaling up mental health care and psychosocial support in low-resource settings.

Roadmap to impact

The ‘roadmap to impact' model bridges research and practice, two sides of a coin that have traditionally been rather divided. Unfortunately, the interventions most implemented in practice tend to be those with the least research evidence, and the findings from research have had limited impact on practice (Tol et al., Reference Tol, Patel, Tomlinson, Baingana, Galappatti, Panter-Brick, Silove, Sondorp, Wessells and van Ommeren2011). Therefore, the roadmap connects the research space with the intervention space, to create a pathway from evidence to practice with measurable indicators along the trajectory (see Fig. 1). First, the research space involves intervention-level research that establishes relevance and effectiveness and expands the evidence-base for task-shifted care. Second, the implementation space involves system-level work that transfers meaningful evidence-based interventions to large-scale impact (Jordans et al., Reference Jordans, van den Broek, Brown, Coetzee, Ellermeijer, Hartog, Steen and Miller2018). In the implementation space, quality and feasibility criteria need monitoring, beginning with benchmarks established in the research phase that are continuously refined through real-world delivery. We will explain the roadmap by describing the four criteria that need to be met to successfully scale, and we provide guidance, by way of example frameworks, on how each criterion can be evaluated.

Fig. 1. Roadmap to impact – research and implementation trajectory to achieve scale. Note: This figure has been adapted from the version published in Jordans et al. (Reference Jordans, van den Broek, Brown, Coetzee, Ellermeijer, Hartog, Steen and Miller2018).

The research space

The starting point for any intervention is demonstrating its relevance (criterion #1) within a given LMIC setting. Given the growing diversity of potential mental health and psychosocial support interventions that vary by treatment duration, training and supervision requirements, and target group or condition, it is important to avoid a one-size-fits-all approach to services. Relevance needs to precede effectiveness, because an effective intervention is unlikely to receive buy-in for scaling if there is not a recognisable benefit at the community and policy levels. Relevance can be demonstrated by two indicators: population need and fit with the culture and context (see Table 1). (a) Population need is the degree to which the intervention addresses actual and current mental health needs within the target population. The Assessing Mental Health and Psychosocial Needs and Resources: Toolkit for Humanitarian Settings is an example framework that provides resources for determining what would be relevant in a particular context, but the advised approaches have not been systematically used in low-resource settings outside of humanitarian context (WHO and UNHCR 2012). There needs to be a clear demonstration of need, as well as mapping of current services to be sure that the proposed intervention does not duplicate existing resources. Likelihood for scaling will be greatest when a gap is being filled. (b) The second indicator is the ‘fit’ or compatibility with culture and context. Ensuring a match between context and intervention has demonstrated to result in superior treatment outcome (Chowdhary et al., Reference Chowdhary, Jotheeswaran, Nadkarni, Hollon, King, Jordans, Rahman, Verdeli, Araya and Patel2014), though this association is not undisputed (Cuijpers et al., Reference Cuijpers, Karyotaki, Reijnders, Purgato and Barbui2018). An example framework for adapting psychological interventions to the culture and context addresses cultural concepts of distress, treatment components and treatment delivery (Heim and Kohrt, Reference Heim and Kohrt2019). An argument has been made that being culturally appropriate is not enough, but rather that successful interventions need to be ‘culturally compelling’ in that the intervention taps into key motivations or drivers for individual, family and community roles within a given culture or context (Panter-Brick et al., Reference Panter-Brick, Clarke, Lomas, Pinder and Lindsay2006), i.e. that interventions address ‘what matters most’ for both beneficiaries and providers (Kohrt et al., Reference Kohrt, Turner, Rai, Bhardwaj, Sikkema, Adelekun, Dhakal, Luitel, Lund, Patel and Jordans2020b).

Table 1. Domains and indicators for roadmap to impact

Note. The darker shaded cells represent the primary application of the indicator within the roadmap, whereas the lighter shaded cells represent the secondary application of the indicators within the roadmap.

Only once the criterion of relevance is met, can one move towards the criterion of effectiveness (criterion #2). The study of the effectiveness of mental health services has been a pillar for psychological treatment research for several decades. Effectiveness studies have exponentially increased in LMIC resulting in a large body of trials in LMIC (Patel et al., Reference Patel, Saxena, Lund, Thornicroft, Baingana, Bolton, Chisholm, Collins, Cooper, Eaton, Hermann, Herzallah, Huang, Jordans, Kleinman, Medina Mora, Morgan, Niaz, Omigbodun, Prince, Rahman, Saraceno, Sarkar, De Silva, Singh, Stein, Sunkel and Unutzer2018). (a) The first indicator of effectiveness is a positive effect size for the mental health outcome of interest. In an umbrella review of 129 primary studies from ten meta-analyses, representing 22 623 participants, Barbui et al. (Reference Barbui, Purgato, Abdulmalik, Acarturk, Eaton, Gastaldon, Gureje, Hanlon, Jordans, Lund, Nosè, Ostuzzi, Papola, Tedeschi, Tol, Turrini, Patel and Thornicroft2020) demonstrate that there is robust evidence for psychosocial interventions for adults with depression, schizophrenia, as well as PTSD in humanitarian settings. For children, the evidence base is much smaller with suggestive evidence only for children with disruptive behaviour or with PTSD in humanitarian settings (Barbui et al., Reference Barbui, Purgato, Abdulmalik, Acarturk, Eaton, Gastaldon, Gureje, Hanlon, Jordans, Lund, Nosè, Ostuzzi, Papola, Tedeschi, Tol, Turrini, Patel and Thornicroft2020). This means that evidence is accruing, but also that there is still some distance to go for a package of interventions with a solid evidence base, covering multiple mental health conditions and a wider age range. Consequently, there is a need for more trials demonstrating effectiveness. For policy makers, cost-effectiveness needs to assessed for interventions for which the current evidence-base is still weak, especially mental health promotion and prevention programmes, and child and adolescent treatments. An example framework to establishing the evidence is a five-step research process used for the evaluation of WHO's newly developed interventions, which involves formative research, a feasibility trial combined with a qualitative process evaluation, and a definitive trial combined with a qualitative evaluation (Bryant et al., Reference Bryant, Schafer, Dawson, Anjuri, Mulili, Ndogoni, Koyiet, Sijbrandij, Ulate and Shehadeh2017). (b) The second indicator of effectiveness is a measured mechanism of action. Besides RCTs demonstrating effectiveness, there is also a need for greater attention to evaluating hypothesised mechanisms of action and to conduct dismantling studies (i.e. measuring separate pieces of interventions) to better understand the active ingredients determining how and why interventions are effective. Mechanisms of action can be measured with self-report scales, but it is preferable to use behavioural observation or capitalise on behavioural data that can, for example, be acquired through passive sensing of mobile digital devices, such as smartphones. With a better understanding of underlying mechanisms of change, one can adjust delivery of the interventions to maximise the benefits of interventions.

The implementation space

For those interventions that do meet both of the aforementioned criteria (relevance and effectiveness), the next challenge is how such interventions can be implemented at scale. We now enter the implementation space in the model. This domain of study, implementation science, is gaining momentum in LMIC (Means et al., Reference Means, Kemp, Gwayi-Chore, Gimbel, Soi, Sherr, Wagenaar, Wasserheit and Weiner2020), and pertains to the study of how evidence-based interventions can be implemented to have the intended results among a large proportion of people in need of care. Attention to the relevance of an intervention can be accomplished through ongoing engagement with service users and community advisory boards, and routine monitoring of client outcomes can shed light on if and for whom an intervention is working. However, the emphasis in the implementation space is on criteria along two axes: (1) quality of care, and (2) feasibility. These two criteria are crucial to monitor the implementation of an evidence-based intervention at scale and understand why an intervention may not be working in a specific setting or with a specific population.

Therefore, in order to assure that the results demonstrated in a well-controlled trial are also achieved in everyday practice, we propose systematic assessment of quality of care (criterion #3), which we have operationalised using three minimum indicators – (a) adherence, (b) competence and (c) attendance. These are defined as the extent to which a service provider has the knowledge and skill required to deliver a treatment to the standard needed for it to achieve its expected effects (competence) and the extent to which a psychological treatment was delivered well enough for it to achieve its expected effects (adherence) (Fairburn and Cooper, Reference Fairburn and Cooper2011). In addition, participants need to receive enough of the intended content (attendance). With this minimum set of indicators, we argue, a programme can assess quality of care at scale; and if adequate levels of competence, adherence and attendance have been obtained, the positive client-level outcome, as demonstrated in the research space, can be assumed, rather than needing to always be measured, which is typically not feasible at scale. For example, a service provider can be highly competent, however, if implementing without adequate adherence to an intervention protocol, then we cannot assume a positive outcome. Similarly, if a provider is meticulously following the intervention protocol, but does not have the core therapeutic skills and competencies, then again positive outcomes cannot be assumed. The same goes for service providers who demonstrate high levels of competence and adherence but who are working in a context where participants are unable to consistently attend and participate in care. If, however, adherence, competence and attendance are all adequate, we assume that we can rely on previously established research findings.

This thinking is commonly advocated by proponents of empirically supported treatments, suggesting that the key to transporting effectiveness findings in everyday clinical settings is ensuring high levels of therapist competence and adherence (Collyer et al., Reference Collyer, Eisler and Woolgar2020). However, for this approach to work in LMIC when working with non-specialists, there is a need for each of these indicators to be validated or benchmarked, such that we know what level of adherence, competence and attendance needs to be obtained to substantiate an assumption of effectiveness. Validation of competence indicators is currently underway in an effort to guide the scaling of psychological treatments by the WHO and partners (Kohrt et al., Reference Kohrt, Schafer, Willhoite, van't Hof, Pedersen, Watts, Ottman, Carswell and van Ommeren2020a). This is based on prior work done to develop new tools that allow for the assessment of competencies that are common across all mental health interventions. These new tools, ENhancing Assessment of Common Therapeutic Factors (ENACT) for adults (Kohrt et al., Reference Kohrt, Jordans, Rai, Shrestha, Luitel, Ramaiya, Singla and Patel2015a), and Working with children – Assessment of Competencies Tool (WeACT) for children (Jordans et al., Reference Jordans, Coetzee, Steen, Koppenol-Gonzalez, Galayini, Diab and Kohrtunder review) have been developed specifically for feasible use with non-specialists in low-resource settings. In the original tool development studies in Nepal and Gaza, we demonstrated that such competency assessment, using standardised role-plays and life observations, captures changes before and after training, and can be used by multiple raters with sufficient reliability (Kohrt et al., Reference Kohrt, Ramaiya, Rai, Bhardwaj and Jordans2015b; Jordans et al., Reference Jordans, Coetzee, Steen, Koppenol-Gonzalez, Galayini, Diab and Kohrtunder review). The tools have also since been used to evaluate the competencies of mental health service providers in several different low-resource settings (Kohrt et al., Reference Kohrt, Asher, Bhardwaj, Fazel, Jordans, Mutamba, Nadkarni, Pedersen, Singla and Patel2018; Rahman et al., Reference Rahman, Akhtar, Hamdani, Atif, Nazir, Uddin, Nisar, Huma, Maselko and Sikander2019).

Benchmarks for adherence, competence and attendance need to be established initially in the research space. Either through reporting of details related to adherence, competence and attendance in RCTs, or through separate validation studies. However, it is within the implementation space that these benchmarks should be refined because there will be considerably greater variation in these three indicators in real-world settings. For example, the rigorous procedure in RCTs is unlikely to lead to the inclusion of providers with low competency. While in routine practice, there will likely be a range of competency levels among providers in public and private health and service institutions. Similarly, in most trials, procedures are in place to promote attendance, and the same level of effort is beyond the scope of most national health systems. Therefore, adjusted benchmarks will likely emerge through real-world implementation of the intervention. These benchmarks can be especially useful for guiding incremental changes in government implementation strategies to improve existing services.

The monitoring of the three indicators and comparison against minimum benchmarks also allows for quality improvement. Tracking levels of adherence, competence and attendance within any given area of programme provides supervisors with an overview of which indicators fall below the validated thresholds. For example, if a group of service providers consistently scores low on certain competencies or specific treatment components, or in case of a trend in drop-outs, then supervisors can remediate this with more targeted and tailor-made solutions. In turn, we hypothesise that such targeted quality monitoring is more cost-effective than approaches that do not have such a data-driven approach. An example framework that brings together tools and knowledge for the assessment of quality of care is WHO's Ensuring Quality in Psychological Support (EQUIP) program (https://www.who.int/mental_health/emergencies/equip/en/) (Kohrt et al., Reference Kohrt, Schafer, Willhoite, van't Hof, Pedersen, Watts, Ottman, Carswell and van Ommeren2020a).

This brings us to the last axis of the model, feasibility (criterion #4), and the associated indicators – (a) coverage and (b) cost. Because even if interventions are relevant and effective, and quality is maintained, population-level impact is only achieved if a large enough proportion of those for whom the intervention is intended are actually reached. The level of uptake is expressed as contact coverage, and defined by Tanahashi (Reference Tanahashi1988) as the ratio between the number of people who have contacted the service and the size of the target population. Reaching a certain level of coverage needs to be determined for scale-up to be deemed successful. An example framework is the Goldberg-Huxley model, which describes a process of help seeking for people with mental disorder along a set of filters that need to be addressed to maximise coverage (Goldberg and Huxley, Reference Goldberg and Huxley1980). An application of that can be found in the Program for Improving Mental Health Care (PRIME), which evaluated the integration of mental health into primary health care (Lund et al., Reference Lund, Tomlinson, De Silva, Fekadu, Shidhaye, Jordans, Petersen, Bhana, Kigozi, Prince, Thornicroft, Hanlon, Kakuma, McDaid, Saxena, Chisholm, Raja, Kippen-Wood, Honikman, Fairall and Patel2012; Jordans et al., Reference Jordans, Luitel, Kohrt, Baron, Rathod, De Silva, Komproe, Patel and Lund2019b). The programme demonstrated that population-level change in contact coverage was not achieved in some of the settings– even after significant efforts to make services available (Nakku et al., Reference Nakku, Rathod, Garman, Ssebunnya, Kangere, De Silva, Patel, Lund and Kigozi2019; Shidhaye et al., Reference Shidhaye, Baron, Murhar, Rathod, Khan, Singh, Shrivastava, Muke, Shrivastava and Lund2019). Besides such supply-side approach, demand-side drivers will therefore need to be addressed in order to actually achieve significant changes in contact coverage. Community-level awareness raising, stigma-reduction approaches and proactive case detection may be strategies that increase demand (Eaton et al., Reference Eaton, Gureje, Silva, Sheikh, Ekpe, Abdulaziz, Muhammad, Akande, Onukogu, Onyuku, Abdulmalik, Fadahunsi, Nwefoh and Cohen2018; Jordans et al., Reference Jordans, Luitel, Lund and Kohrt2020). Contact coverage can be calculated with the attendance data (see criterion 3c) combined with epidemiological data on the prevalence of the condition the intervention is targeting. Finally, for true scalability of impact, the cost of implementation of the intervention needs to be acceptable for settings with limited resources for mental health services. This means that besides evaluating the cost-effectiveness for interventions, the feasible implementation of interventions at scale will need to include monitoring of costs-per-person against set targets specific for certain settings and population (Chisholm et al., Reference Chisholm, Heslin, Docrat, Nanda, Shidhaye, Upadhaya, Jordans, Abdulmalik, Olayiwola, Gureje, Kizza, Mugisha, Kigozi, Hanlon, Adugna, Sanders, Pretorius, Thornicroft and Patel2017; Chisholm et al., Reference Chisholm, Garman, Breuer, Fekadu, Hanlon, Jordans, Kathree, Kigozi, Luitel and Medhin2020). Although trials can establish incremental cost-effectiveness ratios, the actual implementation costs can vary significantly from what was estimated under trial conditions.

In brief, we argue for monitoring the adherence, competence and attendance of relevant evidence-based interventions against validated thresholds or standards, combined with targets for contact coverage and per-person cost. This forms the minimum set of criteria to guide the process of scaling mental health intervention and achieve population-level impact.

Discussion

To date, myriad implementation frameworks have been developed as descriptive, prescriptive, explanatory, or predictive heuristics for the traditional translation pipeline from efficacy to dissemination and implementation, e.g. Replicating Effective Programs (REP), Consolidated Framework for Implementation Research (CFIR), and Reach, Efficacy, Adoption, Implementation, and Maintenance (RE-AIM). Our proposed model provides added value to these frameworks by identifying a minimum set of specific criteria that planners of mental health services can use to plot where a programme is in along the pathway to scalability. It also provides guidance about how to achieve greater impact, giving example frameworks and tools that can be used in practice for achieving each of the criteria. Moreover, the model can be used to address trade-offs between optimizing impact while taking into account constraints on issues related to relevance, effectiveness, quality and feasibility. Existing implementation science frameworks, especially prescriptive frameworks, can be used to elucidate how to move from one region of the roadmap to impact to another region. Finally, the presented model emphasises practicability by proposing a set of concrete and measurable criteria and indicators that have been tried and tested in LMIC, and at scale it can do so with data that can be entirely obtained from the service delivery agents – therefore not relying on individual-level data from participants to demonstrate impact.

One strength of the model is that the quality of care indicators reduces data collection to a minimum. At the same time, we acknowledge that such data collection requires commitment and resources of mental health care planners. In an effort to maintain quality of services, this is likely a worthwhile investment. Moreover, previous efforts to use routine data collection for mental health services have demonstrated feasibility in several low-resource settings (Jordans et al., Reference Jordans, Chisholm, Semrau, Gurung, Abdulmalik, Ahuja, Mugisha, Mntambo, Kigozi, Petersen, Shidhaye, Upadhaya, Lund, Thornicroft and Gureje2019a). More broadly, successful application of this model is dependent on commitment from governments or international agencies. This entails a policy context that prioritises evidence-based mental health care, and the allocation of resources to implement the services at scale, as well as the monitoring framework consisting of the criteria put forward in this paper. National and global investments will be required to develop the infrastructure for these indicators and technical expertise to manage data collection and interpretation of information. Otherwise, adding indicators without a system of analysis and action would risk detracting energy from already stretched-thin public mental health systems. Consequently, the application of the model should be part of a larger effort of mental health system strengthening.

We propose that for such an approach to be operational, a central repository is established where data are stored and accessible for policymakers and practitioners across the global mental health field. In fact, we envision that the validated quality criteria (benchmarks) can become endorsed inter-agency standards. Any agency scaling up evidence-based treatment would therefore strive to achieve these standards, but also agree to report against them in the data repository. If such data, at aggregate level, are made open access it allows for monitoring of overall scaling efforts across geographic areas, across interventions, across organisations. This could draw upon approaches and lessons learned from similar systems for registering RCTs in a public repository (e.g. ClinicalTrials.gov, ISRCTN), data repositories for specific funders (e.g. NIMH Data Archive), collaborative databases for specific conditions (e.g. Autism Brain Imaging Data Exchange) and evidence-based interventions that are searchable by implementation characteristics (e.g. Research-Tested Intervention Programs (RTIPs) for cancer, rtips.cancer.gov).

There are several limitations to the proposed model. First, while the notion of evidence-based care is increasingly being adopted, one can argue that we are still too far removed from having a solid evidence-base in LMICs. Cuijpers et al. (Reference Cuijpers, Karyotaki, Reijnders, Purgato and Barbui2018), synthesizing the literature on treatments for depression, argue that the effects that have been demonstrated in the literature tend to be over-estimated – provokingly asking whether psychological treatments work at all (Cuijpers et al., Reference Cuijpers, Karyotaki, Reijnders and Ebert2019). More attention to strengthening the evidence base is clearly needed (the research space) – for example, for children and adolescents, as also shown in the review by Barbui et al. (Reference Barbui, Purgato, Abdulmalik, Acarturk, Eaton, Gastaldon, Gureje, Hanlon, Jordans, Lund, Nosè, Ostuzzi, Papola, Tedeschi, Tol, Turrini, Patel and Thornicroft2020). We believe that the research agenda can simultaneously emphasise the study of how evidence-based intervention can be implemented at scale (the implementation place), rather than consecutively. The model provides a framework for guiding that research. Second, one might ask whether the level of standardisation of interventions and quality indicators across participants and across cultural settings is possible. As much as interventions will need to be adjusted to new cultural contexts, without changing any of the key working mechanisms, we propose that the quality indicators equally will need to be adjusted for the different settings, including renewed validation studies. Third, the model does not provide an exhaustive overview of the process and indicators involved in scaling, and as such does not do justice to all the complexities involved in scaling. For example, we have not included indicators assessing the political commitment, or to assess quality of care we acknowledge that attendance alone is an insufficient indicator of dosage, as that will also depend on participants' level of engagement with the intervention. We have aimed to keep a minimum set of indicators that is measurable at scale.

Conclusion

This paper provides a framework to guide the implementation of evidence-based mental health and psychosocial interventions at scale in real-world settings, using a streamlined set of criteria to maximise impact at population level. If adequate quality of implementation of evidence-based treatment is obtained (through the assessment of competence, adherence and attendance criteria relying on validated cut-points), combined with adequate feasibility (through the assessment of cost and coverage criteria using a priori set targets), then this provides a foundation for positive outcomes at scale. Scaling without these minimum standards is unlikely to translate evidence-based research into public health impact to reduce the mental health treatment gap around the world.

Data

No data were used for this article.

Acknowledgements

We want to thank Dr Ken Miller and Dr Joop de Jong for their comments on an earlier draft of this paper, as well as other members from the Research and Development Department, War Child Holland, for contributing to the thinking presented here. We acknowledge Dr Ivan Komproe for his inputs on an early version of the model presented in this paper. BAK and MJDJ receive funding from the US National Institute of Mental Health (R01MH120649).

Financial support

This research received no specific grant from any funding agency, commercial or not-for-profit sectors.

Conflict of interest

None.

References

Barbui, C, Purgato, M, Abdulmalik, J, Acarturk, C, Eaton, J, Gastaldon, C, Gureje, O, Hanlon, C, Jordans, MJD, Lund, C, Nosè, M, Ostuzzi, G, Papola, D, Tedeschi, F, Tol, WA, Turrini, G, Patel, V and Thornicroft, G (2020) Efficacy of psychosocial interventions for mental health outcomes in low-income and middle-income countries: an umbrella review. The Lancet Psychiatry 7, 162172.Google ScholarPubMed
Bryant, RA, Schafer, A, Dawson, KS, Anjuri, D, Mulili, C, Ndogoni, L, Koyiet, P, Sijbrandij, M, Ulate, J and Shehadeh, MH (2017) Effectiveness of a brief behavioural intervention on psychological distress among women with a history of gender-based violence in urban Kenya: a randomised clinical trial. PLoS Medicine 14, e1002371, 116.CrossRefGoogle ScholarPubMed
Chisholm, D, Heslin, M, Docrat, S, Nanda, S, Shidhaye, R, Upadhaya, N, Jordans, MJD, Abdulmalik, J, Olayiwola, S, Gureje, O, Kizza, D, Mugisha, J, Kigozi, F, Hanlon, C, Adugna, M, Sanders, R, Pretorius, C, Thornicroft, G and Patel, V (2017) Scaling-up services for psychosis, depression and epilepsy in sub-Saharan Africa and South Asia: development and application of a mental health systems planning tool (OneHealth). Epidemiology and Psychiatric Sciences 26, 234244.CrossRefGoogle Scholar
Chisholm, D, Garman, EC, Breuer, E, Fekadu, A, Hanlon, C, Jordans, MJD, Kathree, T, Kigozi, F, Luitel, NP and Medhin, G (2020) Health service costs and their association with functional impairment among adults receiving integrated mental health care in five low-and middle-income countries: the PRIME cohort study. Health Policy and Planning 35, 567576.CrossRefGoogle ScholarPubMed
Chowdhary, N, Jotheeswaran, AT, Nadkarni, A, Hollon, SD, King, M, Jordans, MJD, Rahman, A, Verdeli, H, Araya, R and Patel, V (2014) The methods and outcomes of cultural adaptations of psychological treatments for depressive disorders: a systematic review. Psychological Medicine 44, 11311146.CrossRefGoogle ScholarPubMed
Collyer, H, Eisler, I and Woolgar, M (2020) Systematic literature review and meta-analysis of the relationship between adherence, competence and outcome in psychotherapy for children and adolescents. European Child & Adolescent Psychiatry 29, 417431.Google ScholarPubMed
Cuijpers, P, Karyotaki, E, Reijnders, M, Purgato, M and Barbui, C (2018) Psychotherapies for depression in low-and middle-income countries: a meta-analysis. World Psychiatry 17, 90101.Google ScholarPubMed
Cuijpers, P, Karyotaki, E, Reijnders, M and Ebert, DD (2019) Was Eysenck right after all? A reassessment of the effects of psychotherapy for adult depression. Epidemiology and Psychiatric Sciences 28, 2130.Google ScholarPubMed
Dorsey, S, Lucid, L, Martin, P, King, KM, O’Donnell, K, Murray, LK, Wasonga, AI, Itemba, DK, Cohen, JA, Manongi, R and Whetten, K (2020) Effectiveness of task-shifted trauma-focused cognitive behavioral therapy for children who experienced parental death and posttraumatic stress in Kenya and Tanzania: a randomized clinical trial. JAMA Psychiatry 77, 464473.CrossRefGoogle ScholarPubMed
Eaton, J, Gureje, O, Silva, M, Sheikh, TL, Ekpe, E, Abdulaziz, M, Muhammad, A, Akande, Y, Onukogu, U, Onyuku, T, Abdulmalik, J, Fadahunsi, W, Nwefoh, E and Cohen, A (2018) A structured approach to integrating mental health services into primary care: development of the Mental Health Scale Up Nigeria intervention (mhSUN). International Journal of Mental Health Systems 12, 11.CrossRefGoogle Scholar
Fairburn, CG and Cooper, Z (2011) Therapist competence, therapy quality, and therapist training. Behaviour Research and Therapy 49, 373378.CrossRefGoogle ScholarPubMed
Fuhr, DC, Weobong, B, Lazarus, A, Vanobberghen, F, Weiss, HA, Singla, DR, Tabana, H, Afonso, E, De Sa, A, D'Souza, E, Joshi, A, Korgaonkar, P, Krishna, R, Price, LN, Rahman, A and Patel, V (2019) Delivering the Thinking Healthy Programme for perinatal depression through peers: an individually randomised controlled trial in India. The Lancet Psychiatry 6, 115127.CrossRefGoogle ScholarPubMed
Goldberg, D and Huxley, P (1980) Mental Illness in the Community: The Pathway to Psychiatric Care. London: Tavistock.Google Scholar
Heim, E and Kohrt, BA (2019) Cultural adaptation of scalable psychological interventions. Clinical Psychology in Europe 1, 122.CrossRefGoogle Scholar
Jordans, MJD, van den Broek, M, Brown, F, Coetzee, A, Ellermeijer, REC, Hartog, K, Steen, HF and Miller, KE (2018) Supporting children affected by war. Mental Health in Refugee and Post Conflict Populations: Theory, Research and Clinical Practice. A. Nickerson and N. Morina. Amsterdam, Springer.Google Scholar
Jordans, MJD, Chisholm, D, Semrau, M, Gurung, D, Abdulmalik, J, Ahuja, S, Mugisha, J, Mntambo, N, Kigozi, F, Petersen, I, Shidhaye, R, Upadhaya, N, Lund, C, Thornicroft, G and Gureje, O (2019 a) Evaluation of performance and perceived utility of mental healthcare indicators in routine health information systems in five low-and middle-income countries. BJPsych Open 5, e70, 17.Google ScholarPubMed
Jordans, MJD, Luitel, NP, Kohrt, BA, Baron, EC, Rathod, S, De Silva, M, Komproe, I, Patel, V and Lund, C (2019 b) Evaluation of a district mental healthcare plan in a low-resource setting in Nepal: outcomes at community, facility and individual levels. PLoS Medicine 16.CrossRefGoogle Scholar
Jordans, MJD, Luitel, NP, Lund, C and Kohrt, BA (2020) Evaluation of proactive community case detection to increase help seeking for mental health care: a pragmatic randomized controlled trial. Psychiatric services: appi. ps. 201900377.CrossRefGoogle Scholar
Jordans, M, Coetzee, JD, Steen, F, Koppenol-Gonzalez, GV, Galayini, H, Diab, SY and Kohrt, BA (under review) Assessment of therapist competency for child and adolescent psychological treatments and psychosocial services in global mental health: Evaluation of feasibility and reliability of the WeACT tool in Gaza, Palestine.Google Scholar
Kohrt, BA, Jordans, MJD, Rai, S, Shrestha, P, Luitel, NP, Ramaiya, MK, Singla, DR and Patel, V (2015 a) Therapist competence in global mental health: development of the ENhancing Assessment of Common Therapeutic factors (ENACT) rating scale. Behaviour Research and Therapy 69, 1121.CrossRefGoogle ScholarPubMed
Kohrt, BA, Ramaiya, MK, Rai, S, Bhardwaj, A and Jordans, MJD (2015 b) Development of a scoring system for non-specialist ratings of clinical competence in global mental health: a qualitative process evaluation of the Enhancing Assessment of Common Therapeutic Factors (ENACT) scale. Global Mental Health 2, e23, 116.CrossRefGoogle ScholarPubMed
Kohrt, BA, Asher, L, Bhardwaj, A, Fazel, M, Jordans, MJD, Mutamba, BB, Nadkarni, A, Pedersen, GA, Singla, DR and Patel, V (2018) The role of communities in mental health care in low-and middle-income countries: a meta-review of components and competencies. International Journal of Environmental Research and Public Health 15, 1279.CrossRefGoogle ScholarPubMed
Kohrt, BA, Schafer, A, Willhoite, A, van't Hof, E, Pedersen, GA, Watts, S, Ottman, K, Carswell, K and van Ommeren, M (2020 a) Ensuring Quality in Psychological Support (WHO EQUIP): developing a competent global workforce. World Psychiatry 19, 115.CrossRefGoogle ScholarPubMed
Kohrt, BA, Turner, EL, Rai, S, Bhardwaj, A, Sikkema, KJ, Adelekun, A, Dhakal, M, Luitel, NP, Lund, C, Patel, V and Jordans, MJD (2020 b) Reducing mental illness stigma in healthcare settings: proof of concept for a social contact intervention to address what matters most for primary care providers. Social Science & Medicine 250, 112852.CrossRefGoogle ScholarPubMed
Kohrt, BA, El Chammay, R and Dossen, SB (2020) Policy makers’ tough choices for psychological interventions in global mental health: learning from multisite studies. JAMA Psychiatry 77, 452454.CrossRefGoogle ScholarPubMed
Lund, C, Tomlinson, M, De Silva, M, Fekadu, A, Shidhaye, R, Jordans, MJ, Petersen, I, Bhana, A, Kigozi, F, Prince, M, Thornicroft, G, Hanlon, C, Kakuma, R, McDaid, D, Saxena, S, Chisholm, D, Raja, S, Kippen-Wood, S, Honikman, S, Fairall, L and Patel, V (2012) PRIME: a programme to reduce the treatment gap for mental disorders in five low- and middle-income countries. PLoS Medicine 9, e1001359.CrossRefGoogle ScholarPubMed
Means, AR, Kemp, CG, Gwayi-Chore, MC, Gimbel, S, Soi, C, Sherr, K, Wagenaar, BH, Wasserheit, JN and Weiner, BJ (2020) Evaluating and optimizing the consolidated framework for implementation research (CFIR) for use in low-and middle-income countries: a systematic review. Implementation Science 15, 119.Google ScholarPubMed
Nakku, JEM, Rathod, SD, Garman, EC, Ssebunnya, J, Kangere, S, De Silva, M, Patel, V, Lund, C and Kigozi, FN (2019) Evaluation of the impacts of a district-level mental health care plan on contact coverage, detection and individual outcomes in rural Uganda: a mixed methods approach. International Journal of Mental Health Systems 13, 63.CrossRefGoogle ScholarPubMed
Panter-Brick, C, Clarke, SE, Lomas, H, Pinder, M and Lindsay, SW (2006) Culturally compelling strategies for behaviour change: a social ecology model and case study in malaria prevention. Social Science & Medicine 62, 28102825.Google ScholarPubMed
Patel, V, Saxena, S, Lund, C, Thornicroft, G, Baingana, F, Bolton, P, Chisholm, D, Collins, PY, Cooper, JL, Eaton, J, Hermann, H, Herzallah, M, Huang, Y, Jordans, MJD, Kleinman, A, Medina Mora, ME, Morgan, E, Niaz, U, Omigbodun, O, Prince, M, Rahman, A, Saraceno, B, Sarkar, K, De Silva, M, Singh, I, Stein, J, Sunkel, C and Unutzer, J (2018) The Lancet Commission on global mental health and sustainable development. The Lancet 392, 15531598.Google ScholarPubMed
Purgato, M, Gross, AL, Betancourt, T, Bolton, P, Bonetto, C, Gastaldon, C, Gordon, J, O'Callaghan, P, Papola, D, Peltonen, K, Punamaki, RL, Richards, J, Staples, JK, Unterhitzenberger, J, van Ommeren, MH, de Jong, JTVM, Jordans, MJD, Tol, WA and Barbui, C (2018) Focused psychosocial interventions for children in low-resource humanitarian settings: a systematic review and individual participant data meta-analysis. The Lancet Global Health 6, e390e400.CrossRefGoogle ScholarPubMed
Rahman, A, Akhtar, P, Hamdani, SU, Atif, N, Nazir, H, Uddin, I, Nisar, A, Huma, Z, Maselko, J and Sikander, S (2019) Using technology to scale-up training and supervision of community health workers in the psychosocial management of perinatal depression: a non-inferiority, randomized controlled trial. Global Mental Health 6, e8, 112.CrossRefGoogle ScholarPubMed
Shidhaye, R, Baron, E, Murhar, V, Rathod, S, Khan, A, Singh, A, Shrivastava, S, Muke, S, Shrivastava, R and Lund, C (2019) Community, facility and individual level impact of integrating mental health screening and treatment into the primary healthcare system in Sehore district, Madhya Pradesh, India. BMJ Global Health 4, e001344.Google ScholarPubMed
Sikander, S, Ahmad, I, Atif, N, Zaidi, A, Vanobberghen, F, Weiss, HA, Nisar, A, Tabana, H, Ain, QU, Bibi, A, Bibi, T, Liaqat, R, Sharif, M, Zulfiqar, S, Fuhr, DC, Price, LN, Patel, V and Rahman, A (2019) Delivering the Thinking Healthy Programme for perinatal depression through volunteer peers: a cluster randomised controlled trial in Pakistan. The Lancet Psychiatry 6, 128139.Google ScholarPubMed
Singla, DR, Kohrt, BA, Murray, LK, Anand, A, Chorpita, BF and Patel, V (2017) Psychological treatments for the world: lessons from low-and middle-income countries. Annual Review of Clinical Psychology 13, 149181.CrossRefGoogle ScholarPubMed
Tanahashi, T (1988) Health service coverage and its evaluation. Bulletin of the World Health Organization 56, 295303.Google Scholar
Tol, WA, Patel, V, Tomlinson, M, Baingana, F, Galappatti, A, Panter-Brick, C, Silove, D, Sondorp, E, Wessells, M and van Ommeren, M (2011) Research priorities for mental health and psychosocial support in humanitarian settings. PLoS Medicine 8, 16.Google ScholarPubMed
van Ginneken, N, Tharyan, P, Lewin, S, Rao, GN, Meera, S, Pian, J, Chandrashekar, S and Patel, V (2013) Non-specialist health worker interventions for the care of mental, neurological and substance-abuse disorders in low- and middle-income countries. Cochrane Database of Systematic Reviews. C. The Cochrane. Chichester, UK, John Wiley & Sons, Ltd.Google Scholar
WHO and UNHCR (2012) Assessing Mental Health and Psychosocial Needs and Resources: Toolkit for Major Humanitarian Settings. Geneva: World Health Organization.Google Scholar
Figure 0

Fig. 1. Roadmap to impact – research and implementation trajectory to achieve scale. Note: This figure has been adapted from the version published in Jordans et al. (2018).

Figure 1

Table 1. Domains and indicators for roadmap to impact