Implementing measurement based care in community mental health: a description of tailored and standardized methods
BMC Research Notes volume 11, Article number: 76 (2018)
Although tailored implementation methods are touted as superior to standardized, few researchers have directly compared the two and little guidance regarding the specific details of each method exist. Our study compares these methods in a dynamic cluster randomized trial seeking to optimize implementation of measurement based care (MBC) for depression in community behavioral health. This specific manuscript provides a detailed, replicable account of the components of each multi-faceted implementation method.
The standardized best practice method includes training, consultation, a clinical guideline, and electronic health record enhancements with the goal to optimize the delivery of MBC with fidelity. Conversely, the tailored, customized and collaborative method is informed by recent implementation science advancements and begins with a needs assessment, followed by tailored training that feeds back barriers data to clinicians, the formation of an implementation team, a clinician-driven clinic-specific guideline, and the use of fidelity data to inform implementation team activities; the goal of the tailored condition is to ensure the intervention and implementation strategies address unique factors of the context. The description of these methods will inform others seeking to implement MBC, as well as those planning to use standardized or tailored implementation methods for interventions beyond behavioral health.
Measurement based care is an evidence-based practice (EBP) that involves the routine use of standardized assessment results (i.e., Patient Health Questionnaire, PHQ-9 for depression severity ) to guide psychotherapy practice , but this practice is used by fewer than 20% of behavioral health clinicians in the United States [3,4,5]. Implementation science has emerged to address this research-to-practice gap. There is mounting evidence that discrete implementation strategies are insufficient, but it is unclear how to best package multifaceted strategies. A Cochrane review indicated variable impact of tailored implementation methods over standardized approaches in healthcare settings . Optimal methods to identify determinants and inform tailoring remain unknown . This manuscript provides a detailed and replicable description of the methods used in an ongoing study that is comparing standardized versus tailored implementation of measurement based care (MBC) on clinician-level (fidelity) and client-level (depression severity) outcomes in community behavioral health settings. See Table 1 for an overview of methods.
Setting and participants
The study protocol is published elsewhere . The main study is a dynamic cluster randomized trial (12 sites randomized to one of four cohorts and then randomized to condition: standardized or tailored) in partnership with the nation’s largest community-based outpatient behavioral health service provider. Participants consist of both clinicians and clients. Clinicians are representative of Masters level counselors (primarily female, Caucasian, less than 50% licensed). Clients are adults with a diagnosis of depression (primary or secondary) and a PHQ-9 score greater than 10, receiving psychotherapy from a participating clinician.
Standardized implementation method
The standardized method was conceptualized as a “best practices” approach to implementation, incorporating six discrete implementation strategies : (1) electronic health record (EHR) enhancements; (2) a needs assessment; (3) training; (4) MBC guidelines; (5) forming a team; and (6) tri-weekly group consultation. The strategies in this condition were led by a clinical psychologist with MBC expertise and took place over a 5-month active implementation period, followed by a 10-month sustainment period.
The PHQ-9 questions were embedded in the EHR to allow clinicians to transfer clients’ scores from paper form to their individual charts. The EHR calculated total PHQ-9 scores and produced symptom trajectories over time. Clinicians used this graph to monitor outcomes and communicate symptom change (or lack thereof) with clients or other providers. Several self-report questions were also embedded into the progress note to collect MBC fidelity data, such as whether clinicians discussed scores with clients, or indicate why the PHQ-9 was not completed. Clinicians were encouraged but not required to answer these questions.
All clinicians completed a battery of self-report measures and a subset attended a focus group during a baseline needs assessment to reveal contextual factors of influence across six domains (i.e., norms & attitudes; structure & process; resources; policies & incentives; networks & linkages; media & change agents) guided by the Framework for Dissemination . Focus group members were identified by clinic administrators using purposeful sampling to achieve extreme variation . The results from the needs assessment were used to characterize clinics but not inform the implementation process, except that several measures were used to prioritize clinicians for attending the consultation meetings (see below).
Clinicians attended a 4-h training 1 month after the needs assessment. The primary goal was to introduce clinicians to MBC and the PHQ-9 and build foundational knowledge and skill regarding the three core components: (1) administer the PHQ-9, (2) review score graphs, and (3) discuss scores with the client. The training covered: MBC introduction and scientific support; core components of MBC; PHQ-9 scientific support; and clinical tips for addressing lack of progress. Active learning strategies were included in the training guided by adult learning principles [12, 13]. The trainer provided didactic content, modeled how to administer the PHQ-9, engaged clinicians in practicing MBC core components and offered immediate feedback, and showed videos on how to use MBC with challenging clients. Clinicians were invited to ask questions and participate in group discussion. Clinicians earned four CEUs and the clinic received equivalent financial incentives to cover productivity for time spent in training.
The recommended guideline to implement MBC was discussed during the initial training and throughout the consultation meetings. This guideline was informed by the available literature (e.g., ) stating that clinicians ought to administer, review, and discuss the PHQ-9 with depressed adult clients in the beginning of every session. This guideline was presented to clinicians as a strong recommendation not a requirement.
Building consultation teams
Although all clinicians were invited to join the consultation meetings, opinion leaders (i.e., individuals who hold influence in a social network) and champions (i.e., individuals who vocally support an innovation), as determined by self-report measures at baseline (i.e., Sociometric Survey  and Opinion Leadership Scale ), were invited to attend the consultation meetings and their schedules were prioritized, noting that they play a pivotal role in their clinic.
Consultation team meetings
It is well-documented that training alone is insufficient for changing behavior and consultation is needed to increase skill and support implementation . The primary goal of the consultation meetings was to facilitate clinician use of MBC with fidelity to the core components. Each meeting was 60 min and was led by the external consultant via video conference. The consultant was proscribed from focusing on problem solving contextual implementation barriers. On average, six consultation meetings were held in each participating clinic during the 5-month active implementation period. For each meeting, two clinicians presented examples of clients with whom they tried but struggled to implement MBC. A standardized case consultation form was completed prior to the meeting to provide the consultant with context and specific questions (see Additional file 1). Active learning strategies (e.g., group discussion, modeling, practice + feedback) were used to engage clinicians. Clinicians received 1 CEU for each meeting.
Tailored implementation method
The tailored implementation method was conceptualized as “customized and collaborative” and it was informed by the Dynamic Sustainability Framework , as well as the Dynamic Adaptation Process , both of which acknowledge that evidence based practices may require adaptation to best fit the context and implementation strategies selected by a diverse group of stakeholders to target emerging barriers may optimize sustainment. The tailored method included the same six implementation strategies as the standardized method to control for dose/time and resources. However, tailoring occurred in the initial training, in the MBC guideline development, and by the implementation teams.
Similar to the standardized method, all clinicians completed a battery of self-report measures, and a subset of clinicians participated in focus groups. However, director interviews and clinic tours were also conducted by the research team as part of the rapid ethnography  to understand the unique clinic contexts. Needs assessment results were used to tailor the initial training and the data was fed back to clinicians to reveal barriers (e.g., insufficient administrative support, lack of integration with current workflow) and facilitators (e.g., leadership support, prior experience with MBC) specific to their site.
Tailored MBC training was modeled after the same 4-h, expert-led, didactic workshop format as the standardized condition, but training in the tailored condition was customized to address clinic-specific barriers identified in the needs assessment. Training was tailored either by adjusting content to be more relevant to the specific clinic, or altering the structure of the training to include other discussions/examples/activities. For instance, at clinics where clinicians noted that they did not view evidence-based practices as clinically useful or as important as clinical judgment, research evidence was presented to demonstrate that clinicians often overestimate client progress (tailoring of content), or audio recordings of clinicians discussing the clinical utility of the PHQ-9 were incorporated into training (tailoring of structure). Clinic-specific barriers and facilitators were presented at the end of training in the tailored, but not the standardized, condition. The trainings in each condition were conducted by different psychologists to avoid contamination.
Forming implementation teams
In order to maximize diversity in team composition and their ultimate impact, implementation teams were formed using a combination of approaches (i.e., network analysis using sociometric survey, attitudinal measures, self-nominations) to identify key individuals to drive and support the MBC implementation. Implementation teams consisted of five to eight members typically including clinic director(s), clinicians, and the study PI (who served as facilitator). Team composition reflected membership from different tiers (e.g., director, team leader, clinician, and office professional) and different teams within each clinic. Social network analysis was used to identify two opinion leaders from each clinic’s baseline advice networks; specifically, opinion leaders who had the highest in-degree centrality (a measure of influence; ) and spanned the network. Two attitudinal measures  were used to identify two clinicians who strongly endorsed positive attitudes toward MBC to serve as champions on the team. Teams were charged with advancing MBC through the phases of implementation, communicating with other stakeholders, engaging in data-based decision making, increasing buy-in, problem-solving, and identifying barriers and finding solutions, for instance .
Implementation teams were given the choice to tailor the MBC guideline at their clinic rather than follow the standardized recommendation. Some examples of guideline tailoring include the expansion of PHQ-9 administration to clients as young as twelve and to clients without depression. One clinic deemed the first week of each month “assessment week” and focused on giving the PHQ-9 to every client during that week only. Another clinic gave each clinician a list of ten clients to focus PHQ-9 administration, with the intention of allowing clinicians to become comfortable with the PHQ-9 before expanding use.
Implementation team meetings
The primary task of implementation teams was to choose strategies to support MBC use during their tri-weekly meetings and enact them over the course of 5 months. While the specific activities of each team differed based on the barriers identified and strategies chosen, there were three common components across teams. First, each team assigned positions of chair, secretary, and evaluation specialist. The chair generated an agenda and led meetings; the secretary took notes and distributed to other members of the team; and the evaluation specialist received, reviewed and presented penetration data (discussed below). Several teams also opted to create a communication specialist position who was responsible for communicating decisions made by the team with others at the clinic.
Second, teams received data regarding their clinic-specific barriers and facilitators from the needs assessment. Teams also had the option to prioritize which barriers to address using a conjoint analysis approach . Teams who opted to prioritize barriers were given a workspace divided into four quadrants (i.e., low versus high on both dimensions) representing feasibility and importance of addressing clinic-specific barriers. Through collaborative discussion, teams sorted barriers into one of the four quadrants with the intention of first addressing barriers that were deemed of high feasibility and high importance.
Third, although in both conditions objective and clinician self-reported MBC fidelity data was collected from the EHR, this data was consolidated into performance summary reports and disseminated to implementation team members by a research associate 1 day prior to their meetings to guide their work. Data reflected discrete reporting periods between team meetings to allow for an assessment of the team’s impact on MBC. Each team meeting included a discussion of the data report to ensure its accuracy, inform changes to future reports if desired, and guide implementation strategies.
The standardized and tailored methods described in this manuscript include some of the most commonly used implementation strategies such as training, consultation, technology enhancements, audit and feedback, network interventions, and a guideline. However, it is unknown if this specific set of strategies (tailored or standardized) are needed to optimize implementation and sustainment. Also, despite efforts to limit tailoring in the standardized condition, the use of active learning strategies in the training, for example, means that the trainer was likely tailoring content based on barriers identified in the moment versus those identified a priori. Our team is working to compare the differential impact of these methods on clinician and client outcomes in 12 clinics of the United States’ largest behavioral health service provider.
electronic health record
measurement based care
Patient Health Questionnaire
Kroenke K, Spitzer RL, Williams JB. The PHQ-9: validity of a brief depression severity measure. J Gen Intern Med. 2001;16(9):606–13.
Scott K, Lewis CC. Using measurement-based care to enhance any treatment. Cogn Behav Pract. 2015;22(1):49–59. https://doi.org/10.1016/j.cbpra.2014.01.010.
Zimmerman M, McGlinchey JB. Why don’t psychiatrists use scales to measure outcome when treating depressed patients? J Clin Psychiatry. 2008;69(12):1916–9.
Hatfield DR, Ogles BM. Why some clinicians use outcome measures and others do not. Adm Policy Ment Health. 2007;34(3):283–91. https://doi.org/10.1007/s10488-006-0110-y.
Jensen-Doss A, Haimes EMB, Smith AM, Lyon AR, Lewis CC, Stanick CF, et al. Monitoring treatment progress and providing feedback is viewed favorably but rarely used in practice. Adm Policy Ment Health. 2016;45:1–14. https://doi.org/10.1007/s10488-016-0763-0.
Baker R, Camosso-Stefinovic J, Gillies C, Shaw EJ, Cheater F, Flottorp S, et al. Tailored interventions to address determinants of practice. Cochrane Database Syst Rev. 2015. https://doi.org/10.1002/14651858.cd005470.pub3.
Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Serv Res. 2017;44(2):177–94. https://doi.org/10.1007/s11414-015-9475-6.
Lewis CC, Scott K, Marti CN, Marriott BR, Kroenke K, Putz JW, et al. Implementing measurement-based care (iMBC) for depression in community mental health: a dynamic cluster randomized trial study protocol. Implement Sci. 2015;10:127. https://doi.org/10.1186/s13012-015-0313-2.
Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10:21. https://doi.org/10.1186/s13012-015-0209-1.
Mendel P, Meredith L, Schoenbaum M, Sherbourne C, Wells K. Interventions in organizational and community context: a framework for building evidence on dissemination and implementation in health services research. Adm Policy Ment Health. 2008;35:21–37. https://doi.org/10.1007/s10488-007-0144-9.
Palinkas LA, Horwitz SM, Green CA, Wisdom JP, Duan N, Hoagwood K. Purposeful sampling for qualitative data collection and analysis in mixed method implementation research. Adm Policy Ment Health. 2015;42(5):533–44. https://doi.org/10.1007/s10488-013-0528-y.
Beidas RS, Kendall PC. Training therapists in evidence-based practice: a critical review of studies from a systems-contextual perspective. Clin Psychol. 2010;17(1):1–30. https://doi.org/10.1111/j.1468-2850.2009.01187.x.
Herschell AD, Kolko DJ, Baumann BL, Davis AC. The role of therapist training in the implementation of psychosocial treatments: a review and critique with recommendations. Clin Psychol Rev. 2010;30(4):448–66. https://doi.org/10.1016/j.cpr.2010.02.005.
Lambert MJ, Whipple JL, Hwinks EJ, Vermeersch DA, Nielsen SL, Smart DW. Is it time for clinicians to routinely track patient outcome? A meta-analysis. Clin Psychol. 2003;10(3):288–301. https://doi.org/10.1093/clipsy/bpg025.
Valente TW, Chou CP, Pentz MA. Community coalitions as a system: effects of network change on adoption of evidence-based substance abuse prevention. Am J Public Health. 2007;97(5):880–6. https://doi.org/10.2105/ajph.2005.063644.
Childers TL. Assessment of the psychometric properties of an opinion leadership scale. J Mark Res. 1986;23(2):184–8. https://doi.org/10.2307/3151666.
Beidas RS, Edmunds JM, Marcus SC, Kendall PC. Training and consultation to promote implementation of an empirically supported treatment: a randomized trial. Psychiatr Serv. 2012;63(7):660–5. https://doi.org/10.1176/appi.ps.201100401.
Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. 2013;8:117. https://doi.org/10.1186/1748-5908-8-117.
Aarons GA, Green AE, Palinkas LA, Self-Brown S, Whitaker DJ, Lutzker JR, et al. Dynamic adaptation process to implement an evidence-based child maltreatment intervention. Implement Sci. 2012;7(1):32. https://doi.org/10.1186/1748-5908-7-32.
Norman DA. Rapid ethnography. In: Aldersey-Williams H, Bound J, Coleman R, editors. The methods lab. London: Design for Ageing Network (DAN) Royal College of Art; 1999. p. 24–5.
Valente TW. Social networks and health: models, methods, and applications. Oxford: Oxford University Press; 2010.
Higgins MC, Weiner J, Young L. Implementation teams: a new lever for organizational change. J Organ Behav. 2012;33(3):366–88. https://doi.org/10.1002/job.1773.
All authors (CCL, AP, MRB, KS, BRM, MH, EN, HK) contributed to the design of some aspect of the methodology described in this manuscript. CCL was the principal investigator who led the project and made all final design decisions. CCL drafted the abstract, introduction, and limitations and future directions section, as well as compiled and edited all other sections. AP drafted the standardized implementation method section and reviewed all sections. All other authors (MRB, KS, BRM, MH, EN, HK) drafted a subsection, provided feedback on the full manuscript draft. All authors read and approved the final manuscript.
The authors declare that they have no competing interests.
Availability of data and materials
Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.
Consent for publication
Ethics approval and consent to participate
This study was approved by the Institutional Review Board of Indiana University (IRB00000222, Study # 1407578183). IRB granted a waiver of written documentation of informed consent as “the research presents no more than minimal risk of harm to subjects and involves no procedures for which written consent is normally required outside of the research context”. Participants were informed verbally and in writing about the nature of the study and that their participation was voluntary.
This study was funded by the US National Institute of Mental Health (NIMH, 1R01MH103310). The views expressed are those of the authors and not necessarily those of the NIMH.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
iMBC Consultation form. A standardized case consultation form completed by the clinician prior to the consultation meeting to provide the consultant with context and the specific consultation question.
About this article
Cite this article
Lewis, C.C., Puspitasari, A., Boyd, M.R. et al. Implementing measurement based care in community mental health: a description of tailored and standardized methods. BMC Res Notes 11, 76 (2018). https://doi.org/10.1186/s13104-018-3193-0