Abstract
Traditional mechanisms for rating adherence or fidelity are labor-intensive. We developed and validated a tool to rate adherence to Motivational Enhancement Therapy–-Cognitive Behavioral Treatment (MET-CBT) through anonymous client surveys. The instrument was used to survey clients in 3 methadone programs over 2 waves. Explanatory and Confirmatory Factor Analyses were used to establish construct validity for both MET and CBT. Internal consistency based on Cronbach's alpha was within adequate range (α > 0.70) for all but 2 of the subscales in one of the samples. Consensus between clients’ ratings (rwg(j) scores) were in the range of 0.6 and higher, indicating a moderate to strong degree of agreement among clients’ ratings of the same counselor. These results suggest that client surveys could be used to measure adherence to MET-CBT for quality monitoring that is more objective than counselor self-report and less resource-intensive than supervisor review of taped sessions. However, additional work is needed to develop this scale.
Introduction
The gap between science and standard practice is especially wide in the area of substance abuse treatment.1–3 More and more states are mandating that behavioral health care providers, like other health care providers, use evidence-based practices (EBPs).1,4 However, EBPs that are introduced to practitioners without ensuring correct implementation are limited in their usefulness and effectiveness.5–7 Therapist manuals and one-time workshops by themselves are known to be ineffective in helping practitioners utilize new skills.8–10 Even when new EBPs are implemented successfully, adherence drift can become a serious problem. 1 Fidelity monitoring and feedback can help curb this type of drift, and is frequently used in the context of research studies.
Several models have been described to monitor use of EBPs. The “gold standard” of fidelity monitoring is expert or supervisor review of video- or audio-taped therapy sessions and/or live observation. 11 In the real world, it is not feasible for many agencies to use these costly and time-consuming measures to monitor quality of treatment, 12 they are more likely to rely on counselors’ self-reported behavior. However, counselors’ self-reported fidelity to new therapies does not correlate well to actual proficiency or use of the skills.13–15 Client ratings may better fit the need for low-burden assessments of EBP adherence, but there is insufficient evidence of their validity and reliability. 16
In this paper, we describe an innovative fidelity monitoring technique: the use of an anonymous, self-administered form to be completed by clients, and the development and initial validation of this fidelity adherence measure. In particular, we developed an instrument for clients to rate their counselors’ use of Motivational Enhancement Therapy and Cognitive Behavioral Therapy (MET-CBT). Both MET and CBT are empirically validated addiction treatment methods,4,11,17,18 but counselors’ adherence to these models is difficult to measure. Indeed, randomized controlled studies conducted to show the efficacy of motivational techniques have often lacked measurement of counselors’ use of the skills involved.19–21 The purpose of this article is to describe the psychometric properties of a client-rated fidelity measure of MET-CBT.
The Implementation Study context
The development of the MET-CBT client-rated fidelity adherence measure was part of a project testing a model of EBP dissemination. The study was funded by the National Institute on Drug Abuse (NIDA) and conducted from August 1, 2005 to July 29, 2008, in three methadone clinics which are all part of a single large urban addiction treatment agency. The study protocol was approved by the Institutional Review Board (IRB) of the Connecticut Department of Mental Health and Addiction Services, where the first three authors serve in the Research Division. The featured treatment was a blended model of (1) MET, which is especially useful in improving clients’ engagement and motivation to change their substance use behaviors; 22 and (2) CBT, which gives clients the needed skills to carry out these changes and to address problems that lead to substance abuse. 23 The project employed on-site training of all counseling staff and supervisors. Supervisors received additional training in how to reinforce MET-CBT skills. In addition several accepted implementation techniques were applied. These included attending to barriers to organizational change; involving all levels of staff in the change process; adaptation of the EBPs to the local setting's procedures; and training supervisors to provide regular feedback to counselors on adherence and skillfulness. Moreover, a member of the research team, termed an ‘implementation shepherd’ was identified to facilitate monthly advisory agency staff meetings during which barriers to implementation were resolved.
Method
Development of the Met-Cbt Client-Rated Adherence measure
Initial generation of items for the client-rated fidelity measure was based on review of the MET-CBT literature and clinical experience of two of the authors, both clinical psychologists, and one of whom is an expert trainer in MET-CBT. We drafted client statements that might reflect the principles of MET (eg, employing empathy; rolling with resistance) and CBT (eg, coping with risk; developing refusal skills) from a client's perspective. Items were written without clinical jargon and were meant to reflect the behaviors and attitudes of clinicians who are using MET-CBT skills in counseling sessions. Initially, enough items in each domain were included so that analyses could identify the strongest items for inclusion in a future, shorter version.
The first draft of items was sent to four national MET-CBT experts. Experts were asked to independently rate each MET-CBT client fidelity item on a scale of 1–5. Scale values were defined as: 1 = Definitely omit (neither item nor its reversal correctly describes MET or CBT principles); 2 = Probably should omit (unnecessary, confusing, and/or does not discriminate MET or CBT from other techniques); 3 = Keep but re-word (discriminates MET or CBT fairly well, but not clearly); 4 = Probably can keep (a good item, as it discriminates MET-CBT adequately, fairly clear, but may need minor re-wording); 5 = Keep as written (Very good item, highly related to MET-CBT, discriminates well, and clearly worded). We also asked each expert to nominate 15 items to be omitted and to suggest wording changes that would help to make items clearer. We specifically targeted a 6th grade reading level. We calculated the discrepancy between expert ratings and examined 12 items with averages less than or equal to 3.0. Also, based on theory and expert opinion, we established 6 subscales that seemed to capture the essence of using MET; they were: (1) Client Centered Focus, (2) Strengths Based, Self-Efficacy, (3) Empathy and Acceptance, (4) Avoiding Argumentation, (5) Rolling with Resistance, and (6) Developing Discrepancy. Likewise, we established two theory-driven subscales for CBT: (1) Functional Analysis and (2) Skills Training.
We presented the draft client-rated adherence measure to agency advisory board members. Per advisory board recommendation, we administered a pilot of the survey to 14 group therapy patients at the agency. We then reviewed the pilot responses and revised the measure per client and advisory board members’ feedback. Changes included adding a “Not Applicable/Don't Know” response and making the lay-out more user-friendly. Finally, we obtained approval from the IRB for the anonymous client survey, the length of which was two double-sided pages.
Data Collection procedures
Prior to the staff training in MET-CBT, clients of the affected methadone programs were asked to complete the MET-CBT client-rated adherence scale. We refer to this phase as Wave 1 data collection. Patients eligible for responding to the anonymous survey were those patients in treatment for at least two months and who were seen by an agency counselor at least monthly. Respondents had to be at least 18 years old, able to read in English, and have no conservator. Confidential lists of eligible patients were generated internally by the agency clinical coordinator and distributed to the three program sites. Receptionists at each of the three sites were given a standard protocol and trained on how to ask identified patients to complete the survey and to cross off names from the confidential eligibility lists when the person either completed a survey, or refused to complete a survey, or withdrew from treatment. Moreover, administrative agency receptionists were trained to place the date and the name of the client's primary counselor on the survey before handing it to the client to complete. People who had difficulty reading or seeing the survey were invited to meet with a research assistant for phone or interview administration. The study participants were asked to insert the completed surveys into a locked box that could only be opened by research staff. The receptionists gave clients who completed the surveys a $10 gift certificate to a local store as an incentive for participation. Given the anonymity of the survey and the careful methods of retrieval of surveys, a waiver of informed consent was granted by the IRB.
After the initial MET-CBT training was completed, the data collection procedure was repeated, using the same eligibility criteria for clients as in Wave 1. No effort was made to specifically follow up with Wave 1 respondents or to exclude them. Thus, these samples should be considered separate cross-sections of agency clients. Approximately 6 months later, the Wave 3 data collection was initiated, following the same eligibility criteria and procedures used in Wave 1 and Wave 2.
Participants
A total of 610 participants completed the client-rated MET-CBT adherence measure. Two were removed from the data analysis due to response bias (circling the same response throughout the survey). Another four were removed due to incomplete surveys, where more than two-thirds of the survey items were missing. Thus, a total of 604 clients were included in the final analyses (194 for Wave 1, 205 for Wave 2, and 205 for Wave 3).
Results
Construct validity
Exploratory Factor Analyses (EFAs)
A multi-step approach was adopted to evaluate the construct validity for the client level CBT and MET adherence scales. First, we determined that there were no significant differences between the background information for clients in Wave 1 and Wave 2, and the total scale score did not change. Therefore, to increase the sample size for the exploratory factor analyses (EFA), we combined these two waves of data. Table 1 shows the background information by sample groups. EFA was then employed to evaluate the factor structure using the combined Wave 1 and Wave 2 data (now termed the development sample). The factor solutions from the EFA were verified through Confirmatory Factor Analyses (CFA) using the Wave 3 data (the cross-validation sample).
Background information for the development and cross-validation samples.
Gender was not measured in the development sample.
Since the goal for the EFA is to identify latent variables, principal-axis factor analysis (PAF), also known as common factor analysis, was selected as the factor extraction method. 24 Because we expected the factors to be correlated, we used oblique rotation rather than orthogonal. The SPSS 15.0 procedure FACTOR (SPSS, 2006) was used to perform the EFA. To determine the number of factors to retain, we used the following criteria: (1) eigen values ≥ 1.0; (2) last substantial drop in the scree plot; (3) interpretability of the solution; and (4) minimum of three items per factor.
The Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy and Bartlett's test of sphericity indicated that both the MET and CBT scales were psychometrically adequate for factor analysis (for CBT, KMO = 0.94, Bartlett's χ
2
= 3276.20, df = 91,
For CBT, the initial factor analysis yielded a three-factor solution; however, three items had unclear factor loadings. For example, one item did not load onto any factor, and two items had high factor loadings for more than one factor, which implied that the factors were not distinct for these items. As suggested by Kahn, 24 these items were removed to obtain a clearer factor solution. The factor analysis was re-run to reveal a two-factor solution. The first factor accounted for 55.25% of the total variance, and the second factor added an additional 7.91% variance. Items that loaded onto the first factor reflected the combined concept of “Functional Analysis/Coping with Risk” (eg, “My counselor helps me figure out other things to do with my time instead of using”), and items loading onto the second factor represented the “Developing Life Skills” domain (eg, “In my sessions, I learn how to solve problems by breaking them down into steps”).
The EFA for the MET scale revealed a potential five factor solution. The five-factor solution accounted for 56.29% of the variance, and reflected the following five domains: “Support Self-Efficacy/Elicit Change Talk” (eg, “My counselor and I talk about what I want for my future”); “Avoid Argumentation” (eg, “I feel like I have to defend myself to my counselor [reverse scored]); “Roll with Resistance” (eg, “My counselor gets mad if I don't follow our treatment plan” [reverse scored]); “Client Centered Perspective” (eg, “I don't really think that my counselor understands me”[reverse scored]); and “Express Empathy/ Acceptance” (eg, “My counselor tries to see things from my point of view”).
Confirmatory Factor Analyses (CFA)
To continue examining the construct validity for the MET-CBT subscales, we performed Confirmatory Factor Analyses (CFAs) on the cross-validation sample. These CFAs allowed us to compare the model fit for the factor structures derived from the EFAs and factors derived from a priori theory-based and expert opinion. CFAs models were estimated using the AMOS 4.0 program 25 with a full information maximum likelihood (FIML) solution. The model fit was evaluated using various fit indices. The non-significant chi-square indicated a good overall model fit. Also, we applied the Hu et al 26 standards, which suggest accepting a Tucker-Lewis Index (TLI, also known as a non-normed fit index: NNFI) ≥0.95, and a comparative fit index (CFI) >0.95. Applying the Steiger 27 recommendations, we checked for a root mean square error (RMSEA) of <0.060. With respect to the latter statistic, most researchers consider a RMSEA less than 0.08 as an acceptable model fit, and a value less than 0.06 a very good model fit. 28 Since not all the models are nested models, we also examined the Akaike information criterion (AIC), for which lower scores indicate a better fit when comparing two nonnested models. 29 Factors were free to correlate with each other, but each item was constrained to load on only one factor. Factor variances were fixed at 1.0 and factor loadings were not constrained.
The CFA results suggest that the two-factor model derived from the EFA (χ
2
(76) = 183.42,
For the MET scale, model fit indices for the five-factor model derived from the EFA showed the best fit for the data (χ
2
(517) = 912.80,
Tables 2A and B show the CFA factor loadings for the two-factor model of CBT, and the five-factor model of MET, both of which were derived from the EFA. All factor loadings were statistically significant at the
Confirmatory Factor Analysis of the CBT client-rated fidelity scale.
Confirmatory Factor Analysis of the MET client level fidelity scale.
Means and standard deviations for the final CBT and MET subscales are presented in Table 3. Tables 4A and B show the inter-correlations among the CBT and MET subscales, as well as the total scores for the development and cross-validation samples.
Descriptive Statistics and Cronbach's alphas for the development and cross-validation samples.
Zero-order correlation among MET-CBT subscales: development sample only.
Correlation is significant at the 0.01 level, 2-tailed;
correlation is significant at the 0.05 level, 2-tailed.
Zero-order correlation among MET-EBT subscales: cross-validation sample only.
Correlation is significant at the 0.01 level (2-tailed);
correlation is significant at the 0.05 level (2-tailed).
Internal reliability
Internal consistencies were examined using Cronbach's alpha for the total scores and subscales, based on the models that had the best model fit statistics from the CFAs. Table 4 presents the alphas for the developmental and cross-validation samples. For the developmental sample, Cronbach's alphas were all within an adequate range (α > 0.70).
30
In contrast, two subscales for the cross-validation sample had low alphas: The Roll with Resistance subscale was 0.37, and the Acceptance subscale was 0.57. The large discrepancy of alpha values for the Roll with Resistance subscale and the Acceptance subscale between the development sample (α: RR = 0.72, and AC = 0.71), and cross-validation samples (α: RR = 0.37, and AC = 0.57) could be the result of sampling variation. To test this explanation, we compared characteristics of the sample. Only two background variables showed significant differences between the development and cross-validation samples: age and length of time with this counselor. An ANOVA analysis revealed that age is related to the Acceptance subscale for the cross-validation sample (F (4, 200) = 2.46,
Inter-Rater Reliability(Consensus Scores)
For a client level adherence measure to be considered valid, it is important to demonstrate that there is consensus among different clients’ ratings of the same counselor. Inter-rater reliability was computed using the consensus index (rwg(j)) developed by James, Demaree and Wolf 31 for multiple item scales. A higher consensus score implies a high degree of agreement among clients rating the same counselor. James et al 32 suggest that rwg(j) scores equal to or larger than 0.70 indicate a good level of consensus among different raters on the same target. As shown in Table 5, for most of the CBT and MET subscales and total scores, the rwg(j) scores are higher than 0.70, except for the Rolling with Resistance (RR) and Client Centered (CC) subscales. Even though the RR and CC subscales have lower consensus scores, they remained in the 0.6 range, which implies a moderate degree of agreement among clients’ ratings of the same counselor.
Consensus ratings for client level MET-CBT measurements.
Discussion
This paper reports on the development and initial validation of a client-rated MET-CBT fidelity adherence measure. It is a crucial first step in the development of a cost-effective and consumer-directed fidelity instrument. Results suggest that there is good initial psychometric evidence for a client-rated MET-CBT fidelity measure, with the possible exception of two subscales, and point to the need for further studies of the measure.
For the development sample, Cronbach's alphas were all adequate, α > 0.70. However, there were two subscales for the cross-validation sample for which the alphas were low: (1) the Roll with Resistance subscale was 0.37, and (2) the Acceptance subscale was 0.57. The large discrepancy of alpha values for the Roll with Resistance subscale and the Acceptance subscale between the development sample (α: RR = 0.72, and AC = 0.71), and cross-validation samples (α: RR = 0.37, and AC = 0.57) could be the result of sampling variation, which would not be a concern. However, it is also possible that clients in the cross-validation sample, with their greater experience with their counselors, present a more accurate and troubling view of these subscales’ internal consistency. Therefore, the reliability of the Acceptance and Rolling with Resistance subscales is inconclusive, and should be subject to further testing and development.
We examined the level of consensus, or rwg(j), among clients rating the same counselor on the counselor's MET-CBT skills. For most of the CBT and MET subscales and total scores, the rwg(j) was higher than the acceptable 0.70, except for the Rolling with Resistance and Client Centered subscales. We hypothesize that these two subscales were lower because they tap counselor skills that are more difficult for clients to recognize. Even though the RR and CC subscales had lower consensus scores, they still were in the 0.6 range, which implies a moderate degree of agreement among clients’ ratings of the same counselor.
It must be emphasized that the current study was conducted in a real-world context, ie, a community-based addictions treatment agency, where it was not possible to completely control for external variables. Also, methadone clinics are quite different from other addiction treatment agencies, especially since the average length of treatment is considerably longer. Thus, our results may not be generalizable to other community clinics. However, using methadone clinics, which have better attendance and longer-term involvement of clients, helped us to obtain data from clients who were better informants about their clinicians’ work. Also, the initial psychometrics for this scale are good; further studies should be designed to provide additional psychometric information on the use of a client-rated MET-CBT adherence measure with different client populations.
This study offers some support for the idea of tracking fidelity through consumer surveys, which promises a lower-cost, more objective measurement. Clearly, additional work needs to be done to develop the scale prior to its use for high-stakes purposes such as reimbursement. If successful, we envision coupling validated client-rated adherence measures with newer technology, such as computers with touch screens to regularly receive data from consumers from each agency program. With greater automation, we will be able to develop a system with the ability to insert known valid and reliable EBP items for different centers or programs as training takes place, and to automatically tie both client outcomes (from the management information system) and client-rated adherence ratings with the program and the direct care staff. In this manner, the therapeutic change process will be better understood, ultimately leading to more effective therapeutic techniques and improved consumer outcomes.
Author Contributions
Conceived and designed the experiments: WU, HL, LF, SS. Analysed the data: HL. Wrote the first draft of the manuscript: WU, HL. Contributed to the writing of the manuscript: LF, SS, SHG, KLS, JLK, MO. Agree with the manuscript results and conclusions: All authors. All authors reviewed and approved of the final manuscript.
Funding
This research was supported by a grant, R21DA19781, from the National Institute on Drug Abuse. We are grateful for the expert supervisor training and consultation provided by Richard Fisher, MSW, and for the participation by agency personnel and clients, who gave of their time and feedback to make this project possible.
Competing Interests
JLK received funding to develop national training resources on MET/CBT 5. Other authors disclose no competing interests.
Disclosures and Ethics
As a requirement of publication author(s) have provided to the publisher signed confirmation of compliance with legal and ethical obligations including but not limited to the following: authorship and contributorship, conflicts of interest, privacy and confidentiality and (where applicable) protection of human and animal research subjects. The authors have read and confirmed their agreement with the ICMJE authorship and conflict of interest criteria. The authors have also confirmed that this article is unique and not under consideration or published in any other publication, and that they have permission from rights holders to reproduce any copyrighted material. Any disclosures are made in this section. The external blind peer reviewers report no conflicts of interest.
