The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×

Abstract

Objective:

Evidence-based assessment (EBA) plays a critical role in the delivery of effective treatments. However, little is known about the assessment practices of mental health clinicians who treat youths and the factors that support EBA. The authors examined when, how, and under what conditions clinicians conduct EBA.

Methods:

In two multidisciplinary surveys (combined N=2,575), clinicians reported how frequently they conducted pretreatment, ongoing, and posttreatment assessments and how frequently they used standardized measures in usual care of youths.

Results:

Although clinicians reported frequent pretreatment, ongoing, and posttreatment assessments, use of standardized measures was rare. Clinician and practice setting characteristics predicted standardized measure use, and a lack of practical assessment tools appears to be a barrier to use of standardized measures in EBA of youths.

Conclusions:

Many clinicians conduct assessments during treatment, but more practical measures and clinician training may improve the integration of standardized measures into routine practice.

HIGHLIGHTS

  • Mental health clinicians who treat youths reported frequent assessments during pretreatment, ongoing treatment, and posttreatment but infrequent use of standardized measures.

  • Clinician and contextual factors affected clinicians’ use of standardized measures.

  • Training and practical measures are needed to improve evidence-based assessment in routine mental health services for youths.

Evidence-based assessment (EBA) includes use of reliable, valid, and standardized measures throughout treatment and is fundamental to evidence-based treatment (1, 2). Pretreatment assessment informs treatment selection and planning. Ongoing assessments are associated with faster improvement across settings, treatments, client ages, and presentations (3). Posttreatment assessment assists with evaluation of outcomes and need for additional services. EBA training initiatives to support EBA use are increasing, and a recent report suggests these initiatives may improve EBA use (4).

Unfortunately, we know little about EBA use in usual care to inform these initiatives. Some surveys indicate whether clinicians assess at all and how often they use standardized measures in EBAs; few clinicians (e.g., 29%−37%) conduct assessments (5, 6), and among those who do, few use standardized measures in the assessment process (12%−40%) (5, 79). Very little is known about when clinicians conduct assessments, except for three surveys of usual care: 39% of U.S. clinicians (7) and 12% of Canadian clinicians (8) reported using ongoing assessment, and 65% of U.K. clinicians reported using pretreatment, ongoing, and posttreatment assessments, but only 29% did so systematically (10). More studies of assessment throughout the treatment process (i.e., pretreatment, ongoing treatment, and posttreatment) are needed to understand how the usual practices of mental health clinicians who treat youths align with EBA.

Further, evaluation of EBA facilitators and barriers may suggest which important features to integrate into EBA initiatives. Individual attitudes are hypothesized to predict behavioral intentions, which, in turn, predict behavior (11). Among clinicians, more positive attitudes toward standardized measures (particularly perceptions about practicality, or time and cost of accessing, administering, and scoring standardized measures) are associated with greater use of these measures (7, 10). Among psychologists, assessment training also has been associated with use of standardized measures (5). This may explain why psychologists, with their discipline’s traditional focus on assessment, are more likely to use rating scales than other clinicians (9). Similarly, cognitive-behavioral therapy (CBT) clinicians report measuring outcomes more frequently than insight-oriented clinicians, perhaps because of greater emphasis on symptom monitoring within CBT (5). Having fewer years of experience is also associated with use of standardized measures (5, 7); this too may relate to training, in that graduate programs have enhanced assessment training to meet demand for demonstrated outcomes.

Practice context is also important. Clinicians working in agencies or group settings (5, 7) and those who are reimbursed by Medicaid (5) report use of more assessments. These clinicians may have more resources than other clinicians have or may be required to conduct assessments. Among psychologists, those working primarily with adults are more likely to use standardized assessments (9). Finally, within one EBA initiative, White youths were more likely than those with unknown race and ethnicity to receive ongoing assessments (12).

Although informative, surveys have often focused on clinicians serving adults, may be limited to one discipline, may examine only ongoing assessment (not pre- or posttreatment assessment), or may follow EBA initiatives. To extend our understanding of EBA to clinicians treating youths in usual care, we surveyed two multidisciplinary samples: a national survey of professional guild members and a statewide survey of Medicaid-billing clinicians. We present information on when, how, and under what conditions clinicians conduct assessments. We hypothesized that a psychology discipline (9), a learning theory orientation (5), positive attitudes toward standardized measures (2, 7), and more adult (5, 8) and Medicaid or low-income (5) clientele would predict greater standardized measure use. We also hypothesized that a private practice setting (7), more clinical experience (7), and more clientele from racial-ethnic minority groups (12) would predict less standardized measure use.

Methods

All procedures were approved by the University of Missouri’s institutional review board. We provided consent information in a cover letter. Participants consented by returning the survey. For the national survey (2007–2008), we randomly selected 1,000 clinicians each from the American Psychological Association, American Academy of Child and Adolescent Psychiatry, American Counseling Association, National Association of Social Workers, and American Association of Marriage and Family Therapy membership rosters. For the statewide survey (2008), we identified 3,084 clinicians from a public listing of health service clinicians who had billed Medicaid the previous year. Clinicians received up to five mailings. The national survey’s adjusted response rate (undeliverable N=347) was 62% (N=2,863); 1,520 who provided youth assessments were included. The statewide survey’s adjusted response rate (undeliverable N=364) was 50% (N=1,348), and 1,055 were included.

We developed both surveys by using the tailored design method (13). Clinicians reported their professional characteristics: discipline, year of highest degree, percentage of youths and clients with low income and from racial-ethnic minority groups, employment setting, and percentage of Medicaid-reimbursed cases (statewide survey only). Clinicians reported their primary theoretical orientation with a recent, representative case of anxiety, depression, or disruptive behavior among youths. For assessment practices, in both surveys we asked, “How often do you conduct an assessment of problems or strengths during the following points of treatment? At intake or the beginning of treatment, ongoing throughout treatment (e.g., weekly, monthly), at or near the last session, or well after the last session (e.g., weeks, months later)?” In the national survey, we asked, “How often do you use the following specific assessment procedures when treating children and adolescents? Standardized checklists for child/family symptoms or functioning (paper-and-pencil measure) completed by child, parent, teacher, or others with a standardized scoring (e.g., Child Behavior Checklist [CBCL])?” In the statewide survey, we asked, “How often do you use standardized assessment measures with the children/families you treat (e.g., Child Behavior Checklist [CBCL], Connors)?” All responses used a 5-point scale: 1, never or almost never; 2, rarely; 3, sometimes; 4, often; and 5, all or most of the time. The national survey also included the Attitudes Toward Standardized Assessment (ASA [2]), a 22-item measure of attitudes toward standardized measures with three subscales (i.e., benefit over clinical judgment, psychometric quality, or practicality); higher scores on the ASA indicate more positive attitudes.

We first present descriptive information (mean±SD) for assessments at four time points (pretreatment, ongoing treatment, posttreatment, and follow-up). We conducted one-way repeated-measures analysis of variance with a Greenhouse-Geisser correction to compare means across the four time points and a Bonferroni adjustment of p=0.008 for pairwise comparisons. We also present the percentage of clinicians who conducted assessments “often” or “all or most of the time” at each time point and throughout the treatment process (i.e., pretreatment, ongoing, and posttreatment). Second, we present descriptive statistics and the percentage of clinicians who used standardized measures “often” or “all or most of the time” with youths. Third, we conducted a multiple regression analysis predicting standardized measure use by professional discipline (0, not psychology; 1, psychology); percentage of clients from racial-ethnic minorities, with low income, or with Medicaid (statewide survey only); case-mix age (0, primarily [≥50%] youths; 1, primarily adults); private practice setting (0, not private practice; 1, private practice); learning theory orientation (0, other; 1, CBT); and ASA subscale scores (national survey only).

Results

In the national survey (N=1,520), the mean±SD age of clinicians was 52.7±10.0 years; they had 20.4±10.3 years of clinical experience, and most were women (N=958, 63%). The clinicians were Caucasian (N=1,379, 91%), African American (N=39, 5%), Asian/Pacific Islander (N=39, 5%), Latinx (N=43, 3%), other (N=19, 1%), and Native American (N=10, 1%). Most were in private practice (N=922, 61%). A third (N=482, 32%) endorsed learning theory orientation, and more than half had doctorate degrees (N=831, 55%). Their disciplines were psychology (N=482, 32%), counseling (N=357, 23%), marriage and family therapy (N=329, 22%), social work (N=323, 21%), and psychiatry (N=294, 19%). On average, clinicians reported that 32%±28% of their clients belonged to racial-ethnic minority groups and that 35%±32% of the clients had low income. Clinicians reported often conducting assessments pretreatment (4.8±0.6), during treatment (4.2±0.9), and posttreatment (4.0±1.1) and were rarely doing follow-up (2.0±1.2) (F=2,311.44, df=2.60, 3,025.57, p<0.001; all pairwise comparisons p<0.001). Most clinicians conducted pretreatment (N=1,418, 93%), ongoing treatment (N=1,165, 77%), and posttreatment (N=1,032, 68%) assessments often or all or most of the time, and 11% (N=171) used follow-up assessments often or all or most of the time. About half (N=749, 49%) conducted assessments often or all or most of the time throughout the entire treatment process. Clinicians used standardized measures sometimes (3.2±1.4), and 48% (N=726) used them often or all or most of the time. Having more clients from racial-ethnic minority groups (β=0.13, p=0.002) and higher ASA practicality scores (β=0.37, p<0.001) predicted more frequent use of standardized measures, and serving primarily adult clients (β=−0.11, p=0.003) predicted less frequent use of these measures (Table 1).

TABLE 1. Standardized measure use predicted from national and statewide surveys of multidisciplinary cliniciansa

National (N=1,520)Statewide (N=1.055)
Characteristicβpβp
Psychology discipline.07.062.27<.001
Percentage in case mix
 Racial-ethnic minority.13.002.12<.001
 Low income−.03.505.04.236
 Medicaidb−.11.002
Primarily adult case mix−.11.003−.11.001
Private practice−.06.186−.10.002
Learning theory orientation.04.273.06.087
Years of clinical experience−.03.407.06.080
ASA item
 Benefit over clinical judgmentc−.03.549
 Psychometric qualitiesc.01.756
 Practicalityc.37<.001

aPredictors were determined from multiple regression models. Psychology discipline was coded as 0, not psychology, and 1, psychology. Case mix was coded as 0 for primarily youths and 1 for primarily adults. Other case-mix variables (racial-ethnic minority, low income, and Medicaid) reflect percentages of clinicians’ caseload. Private practice was coded as 0, not in private practice, and 1, in private practice. Learning theory orientation was coded as 0, not learning theory, and 1, learning theory (cognitive/behavioral/cognitive-behavioral). ASA, Attitudes Toward Standardized Assessment; β, standardized regression weight.

bOnly the statewide survey included this item.

cOnly the national survey included this item.

TABLE 1. Standardized measure use predicted from national and statewide surveys of multidisciplinary cliniciansa

Enlarge table

In the statewide survey (N=1,055), clinician age was 47.5±11.6 years; they had 13.9±9.4 years of clinical experience, and most were women (N=749, 71%). Clinicians were Caucasian (N=944, 90%), African American (N=71, 7%), other (N=11, 1%), Latinx (N=10, 1%), Native American (N=9, 1%), and Asian/Pacific Islander (N=6, 1%). More than a third (N=410, 39%) endorsed learning theory orientation. Most had a master’s degree (N=763, 72%), and most were in private practice (N=651, 59%). Disciplines included counseling (N=533, 51%), social work (N=374, 35%), psychology (N=282, 27%), marriage and family therapy (N=156, 15%), and psychiatry (N=8, 1%). On average, the clinicians reported that of their clients 28%±30% belonged to racial-ethnic minority groups, 32%±24% had low income, and 53%±35% were Medicaid insured. Clinicians conducted pretreatment (4.80±0.55), ongoing treatment (4.1±1.0), and posttreatment (4.1±1.2) assessments often and follow-up (1.9±1.0) never or almost never (F=2,507.99, df=2.75, 2,776.31, p<0.001). Pretreatment assessment was more frequent and follow-up assessment less frequent than assessments at the other two time points (p<0.001). Most clinicians conducted pretreatment (N=1,005, 95%), ongoing treatment (N=811, 77%), and posttreatment (N=768, 73%) assessments “often” or all or most of the time; 7% (N=78) conducted assessments at follow-up. More than half (N=582, 55%) conducted assessments often or all or most the time throughout the treatment period. On average, clinicians used standardized measures sometimes (3.1±1.3), and 38% (N=404) used them often or all or most of the time. Psychology discipline (β=0.27, p<0.001) and having more clients from racial-ethnic minority groups (β=0.12, p<0.001) predicted more frequent standardized measure use, and serving primarily adults (β=−0.11, p=0.001), being in a private practice setting (β=−0.10, p=0.002), and serving more Medicaid clients (β=−0.11, p=0.002) predicted less frequent use of these measures (see Table 1).

Discussion and Conclusions

Using two large multidisciplinary surveys, we examined when, how, and under what conditions clinicians involved in usual care conduct EBAs with youths. On average, clinicians reported often assessing before, during, and after treatment. Consistent with EBA, approximately half of the clinicians routinely conducted assessments throughout the three treatment stages. However, consistent with previous studies (5, 79), the clinicians used standardized measures only sometimes, and less than half routinely used standardized measures. This discrepancy suggests that assessment most often involves informal unstandardized methods, and EBA training initiatives are needed to support the routine use of standardized measures among clinicians.

These results point to two drivers of standardized measure use that may inform future EBA initiatives: practical measures and clinician training. In line with previous studies in which clinicians cited lack of practicality as a primary barrier to standardized measure use (7, 10), ASA practicality was the greatest predictor of standardized measure use in the national survey, suggesting that practical measures may support routine use of standardized measures. In the statewide survey, a greater percentage of Medicaid cases was associated with less frequent standardized measure use. The state survey had no mandated assessment requirements (this finding may not hold if EBA is required for Medicaid billing), and our finding may largely reflect that many clinicians in this survey may work in low-resource settings. As such, EBA initiatives should consider free or low-cost and brief measures that can be easily integrated into universal care.

Training to support standardized measure use throughout the treatment process and with specific populations is also needed. Consistent with previous surveys (9), psychology discipline predicted increased use of standardized measures. If greater assessment training in psychology graduate programs is driving these results, EBA training for all disciplines may yield improved attitudes, comfort with, and use of EBA measures. Private practice setting was also associated with less frequent standardized measure use. Given that clinicians in private practice may have less favorable views of standardized measures (7), training that increases positive attitudes toward standardized measures may increase their use in private practice. Consistent with another survey (8), clinicians serving primarily adult clients used standardized measures less frequently with youths. Having a higher proportion of Medicaid clients was also associated with less frequent use of standardized measures; this was contrary to our hypothesis but consistent with recent findings of a quality improvement project (14). Hence, addressing issues specific to assessments (e.g., multi-informant reports) of youths and increasing familiarity with standardized measures for youths from diverse backgrounds both seem important. Of note, contrary to past research (12), having a greater proportion of youths from racial-ethnic minority groups predicted greater standardized measure use in both samples; perhaps clinicians serving more racial-ethnic minority youths are using EBA more often because of the smaller body of literature on treatment of youths from racial-ethnic minority groups (15).

This study had some limitations. Clinicians may have overreported use of EBAs because of social desirability or recall bias. Clinicians who conduct more EBAs may have also been more likely to respond, although our response rates were comparable to or higher than those in previous surveys (59), and we included two samples to increase representativeness (i.e., the Medicaid sample was recruited because guild members may have more resources or be more current on standards than nonguild members). Given the timing of these surveys and the push for greater EBA use in the past decade, standardized measure use may now be higher. A multidisciplinary survey is needed to evaluate current assessment practices. Nevertheless, our results provide a useful data point for understanding assessment practices as they evolve. Our surveys were administered as routine progress monitoring was being recognized as an evidence-based practice and before any large-scale EBA initiatives. As such, our findings may provide a benchmark for gauging the success of EBA initiatives.

Department of Psychological Sciences, University of Missouri, Columbia.
Send correspondence to Dr. Hawley ().

This research was supported in part by National Institute of Mental Health grants R03 MH-077752 and P30 MH-068579.

The authors report no financial relationships with commercial interests.

References

1 Hunsley J, Mash EJ: Evidence-based assessment. Annu Rev Clin Psychol 2007; 3:29–51Crossref, MedlineGoogle Scholar

2 Jensen-Doss A, Hawley KM: Understanding barriers to evidence-based assessment: clinician attitudes toward standardized assessment tools. J Clin Child Adolesc Psychol 2010; 39:885–896Crossref, MedlineGoogle Scholar

3 Scott K, Lewis CC: Using measurement-based care to enhance any treatment. Cognit Behav Pract 2015; 22:49–59Crossref, MedlineGoogle Scholar

4 Hall CL, Moldavsky M, Baldwin L, et al.: The use of routine outcome measures in two child and adolescent mental health services: a completed audit cycle. BMC Psychiatry 2013; 13:270Crossref, MedlineGoogle Scholar

5 Hatfield D, Ogles BM: The use of outcome measures by psychologists in clinical practice. Prof Psychol Res Pr 2004; 35:485–491CrossrefGoogle Scholar

6 Phelps R, Eisman EJ, Kohout J: Psychological practice and managed care: results of the CAPP practitioner survey. Prof Psychol Res Pr 1998; 29:31–36CrossrefGoogle Scholar

7 Jensen-Doss A, Haimes EMB, Smith AM, et al.: Monitoring treatment progress and providing feedback is viewed favorably but rarely used in practice. Adm Policy Ment Health Ment Health Serv Res 2018; 45:48–61Crossref, MedlineGoogle Scholar

8 Ionita G, Fitzpatrick M: Bringing science to clinical practice: a Canadian survey of psychological practice and usage of progress monitoring measures. Can Psychol 2014; 55:187–196CrossrefGoogle Scholar

9 Palmiter DJ Jr: A survey of the assessment practices of child and adolescent clinicians. Am J Orthopsychiatry 2004; 74:122–128Crossref, MedlineGoogle Scholar

10 Johnston C, Gowers S: Routine outcome measurement: a survey of UK child and adolescent mental health services. Child Adolesc Ment Health 2005; 10:133–139Crossref, MedlineGoogle Scholar

11 Ajzen I: The theory of planned behavior. Organ Behav Hum Decis Process 1991; 50:179–211CrossrefGoogle Scholar

12 Edbrooke-Childs JH, Gondek D, Deighton J, et al.: When is sessional monitoring more likely in child and adolescent mental health services? Adm Policy Ment Health Ment Health Serv Res 2016; 43:316–324Crossref, MedlineGoogle Scholar

13 Dillman DA: Mail and Internet Surveys: The Tailored Design Method. New York, Wiley, 2000Google Scholar

14 Liu FF, Cruz RA, Rockhill CM, et al.: Mind the gap: considering disparities in implementing measurement-based care. J Am Acad Child Adolesc Psychiatry 2019; 58:459–461Crossref, MedlineGoogle Scholar

15 Pina AA, Polo AJ, Huey SJ: Evidence-based psychosocial interventions for ethnic minority youth: the 10-year update. J Clin Child Adolesc Psychol 2019; 48:179–202Crossref, MedlineGoogle Scholar