The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×
Published Online:https://doi.org/10.1176/appi.ps.201400206

Abstract

Objective:

Characteristics associated with participation in training in evidence-informed business and clinical practices by 346 outpatient mental health clinics licensed to treat youths in New York State were examined.

Methods:

Clinic characteristics extracted from state administrative data were used as proxies for variables that have been linked with adoption of innovation (extraorganizational factors, agency factors, clinic provider-level profiles, and clinic client-level profiles). Multiple logistic regression models were used to assess the independent effects of theoretical variables on the clinics’ participation in state-supported business and clinical trainings between September 2011 and August 2013 and on the intensity of participation (low or high). Interaction effects between clinic characteristics and outcomes were explored.

Results:

Clinic characteristics were predictive of any participation in trainings but were less useful in predicting intensity of participation. Clinics affiliated with larger (adjusted odds ratio [AOR]=.65, p<.01), more efficient agencies (AOR=.62, p<.05) and clinics that outsourced more clinical services (AOR=.60, p<.001) had lower odds of participating in any business-practice trainings. Participation in business trainings was associated with interaction effects between agency affiliation (hospital or community) and clinical staff capacity. Clinics with more full-time-equivalent clinical staff (AOR=1.52, p<.01) and a higher proportion of clients under age 18 (AOR=1.90, p<.001) had higher odds of participating in any clinical trainings. Participating clinics with larger proportions of youth clients had greater odds of being high adopters of clinical trainings (odds ratio=1.54, p<.01).

Conclusions:

Clinic characteristics associated with uptake of business and clinical training could be used to target state technical assistance efforts.

The treatment of psychiatric disorders is among the five most costly medical conditions among children, constituting $8.9 billion or 9% of total annual medical spending for children (1). Research has identified a range of evidence-based practices for psychiatric conditions among youths (24). An ongoing challenge is the effective implementation of evidence-based practices in real-world settings; it takes 17 years for original research to change practice (5).

Over the past decade, states have led the way in promoting evidence-based practices to improve quality of care and reduce costs. Over 20 states are actively implementing evidence-based psychosocial therapies (6) or medication practices (7). Evidence-based practice implementation can be economically sound, with one study reporting a 56% rate of return on investment (8). The demand for states to implement evidence-based practices is matched by a demand for practical strategies and feasible measurement systems to track the impact of implementation (9,10).

New York State has been a leader in implementation of evidence-based practices within the child public mental health system. Propelled by the September 11, 2001, attacks and the expectation of trauma-related sequelae, New York State established the Child and Adolescent Trauma Treatments and Services Consortium with federal funds (11). The consortium’s success in conducting large-scale training in evidence-based trauma practices led to the establishment of the Evidence Based Treatment Dissemination Center (EBTDC) (12,13). The EBTDC provides in-depth technical assistance for clinicians around specific evidence-based treatments. However, data from the New York State Office of Mental Health (NYS OMH) indicate that providers at 52% of eligible clinics statewide have participated in any EBTDC initiatives, suggesting uneven adoption (14).

To expand training beyond traditional evidence-based treatments, New York State established the Clinic Technical Assistance Center (CTAC) in 2011. CTAC’s goals are to strengthen clinics’ capacity for meeting financial and regulatory challenges resulting from state and federal health care reforms. CTAC is thus focused on two core content areas: business and clinical evidence-informed practices. CTAC provides flexible training and consultation models that range from low-intensity, one-hour lunchtime Webinars to intensive learning collaboratives lasting six to 18 months.

Although substantial resources are invested in technical assistance, responses to such training have not been systematically examined. Nationally, strategies for scaling up evidence-based practices have not been guided by empirical knowledge about best practices for implementation (15), and the promising benefits of evidence-based practices have not been consistently realized (16). This article expands on an earlier article that described the training adoption behavior of clinics that are licensed by the NYS OMH to serve youths (17). In the first two years after the establishment of CTAC, 77% of clinics adopted a median of five trainings. Business-practice and clinical trainings were equally accessed, with lunchtime Webinars being the most popular modality. To date, no study has examined what factors, if any, are associated with clinics’ adoption of training. Understanding these factors has significant implications for state strategies to promote clinic training participation and reduce costs associated with training dropout.

Following a theoretical framework of adoption (18), we examined whether multilevel factors, ranging from extraorganizational to client-level factors, predicted clinics’ responses to training. Because predictors of clinic behavior may differ for various types of training, predictor variables for participation in business-practice and clinical trainings were examined separately. Using available administrative data and CTAC data on training participation, this study examined characteristics associated with any participation in business and clinical training and the intensity of participation among clinics that chose to enroll.

Methods

Population

The study population included the 346 clinics licensed by the NYS OMH to treat youths in New York State. However, key fiscal and structural data of interest were not available for clinics affiliated with state-run facilities (N=17), which are structured differently, so these clinics were removed from our analyses. At these facilities, a higher proportion of clients were youth clients (<18 years old) (68.7%±45.2% versus 39.6%±33.4% at other clinics; t=−3.13, df=325, p<.01) and a higher proportion of youth clients had a serious emotional disturbance (76.3%±21.9% versus 35.0%±27.1% at other clinics; t=–4.77, df=306, p<.001). Thus our sample included the 329 outpatient clinics licensed by New York State to treat youths but not operated by the state.

Notably, clinics in New York State can be structured hierarchically, such that one or more clinics are licensed as part of a parent agency. The effects of clinic “nestedness” on each outcome were tested by using chi square tests. Except for one outcome (intensity of business-practice adoption), nestedness was considered ignorable, and agency-level variables were assumed to equally represent all clinics associated with an agency. For the outcome on intensity of business-practice adoption, no models were built, based on our model building criteria described below; none of the variables examined were associated with intensity of business-practice training participation.

Data Sources

Clinic predictor variables.

Variables describing clinic characteristics that could be predictive of training adoption were extracted from four data sources. The U.S. Department of Health and Human Services Area Health Resources Files (AHRF) (ahrf.hrsa.gov) provided county demographic data. The NYS OMH directory of licensed clinics (bi.omh.ny.gov/bridges/index) provided clinic demographic data. The 2011 NYS OMH Consolidated Fiscal Report (CFR) system (www.omh.ny.gov/omhweb/finance/main.htm) provided annual operational capacity and provider-level profiles. The 2011 NYS OMH Patient Characteristics Survey (PCS) (bi.omh.ny.gov/pcs/index) provided “snapshot” data collected during a one-week period in October 2011 on client populations served by the 329 clinics. Depending on the variable, 3.64% to 9.42% of clinics were missing data from the CFR system and 3.34% to 9.42% were missing data from the PCS. Observations with missing data were dropped from statistical analyses.

Outcome variables.

Attendance data were prospectively tracked by training event (17). Because adoption of an evidence-based practice, defined as the decision to proceed with partial or complete implementation, is a distinct process that precedes implementation (1820), sending clinicians to be trained in an innovation represents an early adoption behavior by clinics. A clinic is counted as a training adopter if any individual provider affiliated with the clinic participates in any CTAC trainings. Among participating clinics, an average of 6.87±5.89 providers per clinic participated in any business-practice trainings; 7.09±5.35 clinicians per clinic participated in any clinical trainings. For each outcome variable (business-practice or clinical training), clinics were categorized in terms of any training adoption (yes or no) as well as by intensity of participation among clinics that participated in at least one training (low or high). Low training adopters were defined by exclusive participation in hour-long Webinars, and high adopters were defined by participation in trainings that required in-person attendance, attendance at multiple events associated with a training, or both.

Measures of Clinic Variables

A theoretical model of innovation adoption developed by Wisdom and colleagues (18) provided a framework for clinic variable selection. Selection of the measures, described below, was guided by a review of constructs associated with four levels of predictors of innovation adoption (19).

Extraorganizational variables.

Region-urbanicity served as a proxy for extraorganizational social networks (regions administratively linked by OMH) and training access (distance to urban locations where trainings are typically held). Clinics were initially categorized by region as downstate (New York City and Long Island) or upstate (Central, Hudson, and Western regions) and as urban or rural on the basis of AHRF county rural-urban continuum codes. Because downstate New York State is only urban, a new variable was used reflecting three categories of region-urbanicity: downstate-urban, upstate-urban, and upstate-rural.

Agency-level variables.

Each NYS OMH–licensed clinic is associated with a parent agency. Relevant agency-level attributes included affiliation, operational size, fiscal efficiency, and operational structure. Affiliation refers to whether agencies are classified as community-based providers or hospitals; it served as a proxy for interorganizational social networks and operational structures. The agency’s annual clinic-related total expenses, used by NYS OMH as a measure of operational size, served as a proxy for a variety of theory-based variables, such as absorptive capacity and resource availability (19,20). The fiscal efficiency of clinics affiliated with an agency is a derived variable defined by NYS OMH as the annual average dollar gain or loss per unit of standardized service provided. This measure took into account the type and length of each service encounter. Operational structure, or the annual proportion of clinical workforce (providers who bill for direct services) to total workforce, was used as a proxy for staffing structure.

Clinic-provider profile variables.

The CFR data provided two measures of clinic operational structure: clinical capacity, measured by the annual total clinical full-time-equivalents (FTEs), and outsourced services, measured by the proportion of all clinical staff whose services were contracted. Clinical capacity and degree of outsourcing are hypothesized to influence clinic investment in workforce training (strategic fit) (21).

Clinic-client profile variables.

The PCS provided information that served as proxies for innovation values fit (22). These variables included the proportion of youth clients; proportion of Medicaid or Medicaid managed care (MMC) visits, estimated by the proportion of youth visits billed to Medicaid or MMC; and proportion of youth clients with a serious emotional disturbance, a measure of clinical severity among youths served.

Statistical Analyses

Analyses were performed by using Stata, version 11.2 (23). Descriptive statistics were used to characterize the clinic population according to the two outcome variables of interest: participation in business-practice trainings and clinical trainings. For each outcome variable, we examined clinic characteristics according to training adoption status (yes or no) and intensity of participation (low or high) among adopters. Chi square and t tests were used to compare categorical and continuous variables, respectively. Distributions of four continuous predictor variables (total expenses, gain or loss per unit of standardized service, total clinical FTEs, and proportion of clinical staff contracted out) were notably skewed; natural log-transformed versions of these variables were also tested for bivariate associations with the outcomes.

Multiple logistic regression models (adjusted odds ratios [AORs]) were used to assess the independent effects of predictor variables on clinic training participation. The same model-building procedure was used in four logistic regression models to predict any participation (yes or no) and intensity of participation (low or high) in each type of training (business and clinical).

The independent variables were entered into the models in the following ordered categories: extraorganizational variables, agency-level variables, clinic-provider profile variables, and clinic-client profile variables. Akaike Information Criterion values upon entry of each variable category into the model were compared to determine whether the information gained compensated for the increased model complexity and potential overfitting. Once the main effects for each model were determined, we explored the interactions of the following combinations of categorical and continuous variables, provided both pairs of variables were included among the main effects: region-urbanicity by total expenses, gain or loss per service unit, total clinical FTEs, and proportion of contracted clinical staff; and clinic affiliation by total expenses, gain or loss per service unit, and clinical capacity. Because none of the hospital-affiliated clinics outsourced any clinical services, interaction effects between clinic affiliation and proportion of services outsourced could not be examined. Once main effects and significant interaction terms were determined, reduced models created by removing statistically nonsignificant variables were tested to yield more parsimonious models with improved fit. For ease of interpretability and between-variable comparisons, all numeric variables were standardized for use in the models. Subsequently, model coefficients were interpreted as the change in odds of training adoption for an increase of one standard deviation in the respective predictor variable. All hypothesis tests were two-sided with alpha set at .05. Model fit of each final, reduced model was assessed with Hosmer and Lemeshow’s goodness-of-fit test by using eight predicted value groups. Specification errors were tested by checking for independent variable correlations with the model error term.

Results

A majority of clinics were located in downstate, urban areas (59%) and were affiliated with community-based agencies (82%) (Table 1). The agencies with which the clinics were affiliated had average clinic-related total expenses of $6.04 million. Clinics operated at an average loss of $48.6 per unit of service provided and with a clinical workforce that represented 68.1% of the clinic’s overall workforce. Clinic-provider profiles indicated an average of 12.8 clinical FTEs on staff. On average, 8.3% of clinical services were provided by staff through outside contracts. Clinic-client profiles indicated that on average, youth clients constituted 39.7% of clients, of whom 49.8% were covered by Medicaid or MMC and 35.0% were classified as having a serious emotional disturbance.

TABLE 1. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old))

VariableN%
Extraorganizational (region-urbanicity)
 Downstate-urban19459
 Upstate-urban9830
 Upstate-rural3711
Agency level
 Affiliation
  Community26982
  Hospital6018
 Total expenses, in millions (M±SD $)6.04±6.91
 Gain or loss per service unit (M±SD $)–48.56±91.70
 Clinical staff as a percentage of total workforce (M±SD %)68.14±16.02
Clinic-provider profile
 Clinical full-time-equivalents (M±SD)12.75±10.11
 Proportion of clinical staff contracted out (M±SD %)8.26±17.43
Clinic-client profile
 Youth clients (M±SD %)39.65±33.35
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SD %)49.78±20.81
 Youth clients with a serious emotional disorder (M±SD %)35.00±27.11

TABLE 1. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old))

Enlarge table

Table 2 reports clinic population characteristics by uptake of business-practice trainings. Compared with nonparticipants, clinics that participated in any business trainings were disproportionately located upstate (χ2=16.65, df=2, p<.001), had fewer total agency clinic-related expenses (t=2.21, df=315, p<.05), had more total clinical staff FTEs (t=−2.78, df=296, p<.01), had a smaller proportion of contracted clinical staff (t=3.09, df=296, p<.01), and served more youths (t=−2.31, df=311, p<.05) and more youths with a serious emotional disturbance (t=−2.36, df=296, p<.05). Among adopters of business-practice trainings, none of the variables examined were associated with intensity of training participation.

TABLE 2. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of business-practice trainings

No uptake (N=119, 36%)Any uptake (N=210, 64%)Any uptake vs. no uptakea
VariableN%N%pAORb95% CIp
Extraorganizational (region-urbanicity)<.001
 Downstate-urban877310751
 Upstate-urban26227234
 Upstate-rural653115
Agency level
 Affiliationns
  Community (reference)927717784
  Hospital27233316.50.18–1.36ns
 Total expenses, in millions (M±SE $)7.19 ±.765.41 ±.44<.05.65.50–.84<.01
 Gain or loss per service unit (M±SE $)–45.1±7.5–50.5±7.0ns.62.41–.94<.05
 Clinical staff as a percentage of total workforce (M±SE %)65.9±1.869.4±1.0ns
Clinic-provider profile
 Clinical full-time-equivalents (FTEs) (M±SE)10.6 ±.913.9 ±.7<.011.33.94–1.88ns
 Proportion of clinical staff contracted out (M±SE %)12.5±2.46.0 ±.9<.01.60.46–.80<.001
Clinic-client profile
 Youth clients (M±SE %)33.9±3.242.9±2.3<.05
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SE %)49.3±2.050.1±1.4ns
 Youth clients with a serious emotional disorder (M±SE %)30.0±2.637.7±2.0<.05
Hospital (vs. community) affiliation × total clinical FTEs4.891.31–18.28<.05

aResults are from a reduced logistic regression model for 287 clinics that removed statistically nonsignificant variables (p≥.05); pseudo R2=.1213, likelihood ratio χ2=44.71, df=6, p<.001

bAdjusted odds ratio

TABLE 2. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of business-practice trainings

Enlarge table

Table 2 also displays results from reduced logistic regression models describing any business-practice training adoption as a linear function of the clinic characteristics. Clinics associated with agencies that had larger total expenses (AOR=.65, 95% confidence interval [CI]=.50–.84), had greater gains per unit of service provided (AOR=.62, CI=.41–94), outsourced more clinical care (AOR=.60, CI=.46–.80), and had lower odds of participating in any business-practice trainings. A statistically significant interaction effect of clinic affiliation and total clinical FTEs on the probability of uptake of any business-practice trainings was observed (AOR=4.89, CI=1.31–18.28). Among clinics with fewer total clinical FTEs, hospital-affiliated clinics were less likely to participate in any business trainings; among clinics with more total clinical FTEs, hospital-affiliated clinics were more likely to participate (Figure 1).

FIGURE 1.

FIGURE 1. Interaction effect of total number of clinical full-time-equivalents (FTEs) and affiliation of parent agency on probability of adoption of business-practice trainings by mental health clinics

Table 3 reports clinic population characteristics by clinical-training participation. Compared with nonparticipants, clinics that participated in any clinical trainings were more likely to be affiliated with agencies that had a larger proportion of clinical workforce (t=−2.08, df=316, p<.05). These clinics also had more total clinical staff FTEs (t=−2.84, df=296, p<.01) and served more youth clients (t=−4.93, df=311, p<.001) and more youth clients who were covered by Medicaid or MMC (t=−3.62, df=316, p<.001). Among participating clinics, only one variable was found to be associated with intensity of clinical-training uptake; on average, the percentage of youth clients was higher at clinics with high versus low participation (54.6%±3.7% versus 39.9%±3.0%, t=−3.14, df=203, p<.01).

TABLE 3. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of clinical trainings

No uptake (N=115, 35%)Any uptake (N=214, 65%)Any uptake vs. no uptakea
VariableN%N%pAOR95% CIp
Extraorganizational (region-urbanicity)ns
 Downstate-urban776711755
 Upstate-urban28247033
 Upstate-rural1092713
Agency level
 Affiliationns
  Community 897718084
  Hospital 26233416
 Total expenses, in millions (M±SE $)5.77 ±.716.18 ±.46ns
 Gain or loss per service unit (M±SE $)–44.2±8.1–50.7±6.7ns
 Clinical staff as a percentage of total workforce (M±SE %)65.6±1.769.5±1.0<.05
Clinic-provider profile
 Clinical full-time-equivalents (M±SE)10.5 ±.813.9 ±.8<.011.521.11–2.08<.01
 Proportion of clinical staff contracted out (M±SE %)9.7±1.97.5±1.2ns
Clinic-client profile
 Youth clients (M±SE %)27.3±2.846.1±2.4<.0011.901.42–2.55<.001
 Visits by youth clients paid for by Medicaid or Medicaid managed care (M±SE %)44.1±2.252.8±1.3<.001
 Youth clients with a serious emotional disorder (M±SE %)31.8±2.636.6±2.0ns

aResults are from a reduced logistic regression model for 294 clinics that removed statistically nonsignificant variables (p≥.05); pseudo R2=.079, likelihood ratio χ2=29.74, df=2, p<.001

bAdjusted odds ratio

TABLE 3. Characteristics of 329 outpatient mental health clinics in New York State that serve youth clients (<18 years old) and association with uptake of clinical trainings

Enlarge table

Reduced logistic regression models described any clinical-training participation as a linear function of the clinic characteristics (Table 3). Clinics with more clinical FTEs (AOR=1.52, CI=1.11–2.08) and a larger proportion of youth clients (AOR=1.90, CI=1.42–2.55) had higher odds of attending any clinical trainings. In the model predicting adoption intensity, clinics with a larger proportion of youth clients (OR=1.54, CI=1.17–2.05) had greater odds of high uptake (data not shown).

Discussion

This study is the first, to our knowledge, to examine data on training participation collected prospectively from a population of state outpatient mental health clinics serving youths. To our knowledge, this is also the first attempt to test a multivariate model of adoption behavior involving multilevel characteristics proposed in current theoretical implementation models (18,20,24,25). Our results indicate that available administrative data can be valuable to states in predicting which clinics will participate in training. However, they were much less useful in terms of predicting the level of clinic participation; administrative data are primarily structural and lack more nuanced information—for example, provider attitudes—or local information—for example, details about the organizational culture or climate—that may influence clinic participation levels.

Various levels of clinic characteristics were found to be associated with adoption of different trainings. Business-practice training adoption was associated with agency (affiliation, size, and efficiency) and clinic-provider profile (outsourced clinical services) characteristics. By contrast, adoption of clinical trainings was associated with clinic-provider profile (clinical capacity) and clinic-client profile (proportion of youth clients) characteristics. These findings suggest that clinics make decisions on the basis of relevant agency, provider (strategic fit [21]), and client (innovation values fit [22]) factors.

We speculate that business-practice trainings were less attractive to clinics affiliated with larger agencies, which likely have in-house capacity for fiscal management or need different business-practice supports compared with clinics affiliated with smaller agencies. Clinics associated with more efficiently run agencies and those that outsource more clinical services were also less likely to participate in business-practice trainings. These agencies are likely to have less financial exposure and thus may be in less need of training in business practices. Our data suggest that clinical capacity may influence participation depending on agency affiliation. Among clinics with larger clinical capacity, hospital-affiliated clinics were more likely to participate in business-practice trainings. By contrast, among clinics with smaller clinical capacity, community-based clinics were more likely to adopt business-practice trainings. These findings suggest that operational structure or financial incentives may differ in different types of settings on the basis of clinical capacity, thus influencing business-practice adoption behavior.

Clinical trainings were more likely to be adopted by clinics with larger clinical capacity, which likely reflects clinics’ capacity to release clinical staff for training. Given that the trainings focused primarily on youth-related issues, it seems likely that clinics with a higher proportion of youth clients would consider the trainings more relevant and would be more likely to participate. In-depth interviews with a stratified random sample of these clinics are underway to test our hypotheses.

Implications for State Systems

Our findings suggest that state efforts to provide incentives or target training efforts should pay attention to specific clinic characteristics available through administrative data. Policy makers should understand factors that influence the types of training and the amount of training that clinics are willing or able to adopt. To help clinics move beyond training adoption to successful implementation, targeting training to address both clinic needs and clinics’ readiness for change is critical (21). Health care reform and new accountability standards make it imperative for states to understand and leverage factors that influence clinic decisions to embrace, implement, and sustain quality improvement efforts.

Limitations

When used as the only data source, administrative data have limited value in predicting adoption of training. It may be helpful to collect and examine data about other theoretical factors that are hypothesized to drive adoption behavior in advance of large-scale rollouts of training initiatives. Information about providers—such as attitudes, organizational context, and baseline competencies about business practices and clinical knowledge—must be integrated with administrative data to help policy makers determine appropriate levels of technical support for successful quality improvement efforts.

In addition, clinics may adopt training not associated with CTAC; we were unable to account for the relationship between adoption of CTAC trainings versus other training. National differences in relationships between clinics, counties, and states may limit the generalizability of our findings outside New York State. Nonetheless, our findings provide support for existing theoretical adoption or implementation models that explicate multilevel influences.

Conclusions

This study represents a first effort to understand multilevel characteristics that influence training adoption within a state outpatient child mental health system. Although this study focused on baseline adoption behaviors (training participation), how adoption translates into actual implementation of effective practices to improve care quality and youth outcomes is a critical question. The answer is likely to inform development of optimal models for rollout of state training in service practices.

Dr. Olin, Mr. Weaver, Dr. Kerker, Dr. Hoagwood, and Dr. Horwitz are with the Department of Child and Adolescent Psychiatry, New York University School of Medicine, New York (e-mail: ). Dr. Kerker is also with the Nathan S. Kline Institute for Psychiatric Research, Orangeburg, New York. Dr. Chor is with the Health and Social Development Program, American Institutes of Research, Chicago. Dr. Duan is with the Division of Biostatistics, Department of Psychiatry, Columbia University, New York. Ms. Clark is with the New York State Office of Mental Health, Albany. Dr. Cleek is with the Silver School of Social Work, New York University, New York.

This study was supported by grants P30 MH090322 and P20 MH085983 from the National Institute of Mental Health and by the Clinic Technical Assistance Center, funded by the New York State Office of Mental Health.

The authors report no financial relationships with commercial interests.

References

1 Mental Disorders Among Most Costly Conditions in Children. Rockville, Md, Agency for Healthcare Research and Quality, 2009. Available at archive.ahrq.gov/news/newsroom/news-and-numbers/042209.htmlGoogle Scholar

2 Burns BJ, Hoagwood KE: Evidence-based practice, part I. a research update. Child and Adolescent Psychiatric Clinics of North America 13:xi–xiii, 2004CrossrefGoogle Scholar

3 Kazdin AE, Weisz JR (eds): Evidence-Based Psychotherapies for Children and Adolescents. New York, Guilford, 2003Google Scholar

4 National Registry of Evidence-based Programs and Practices. Rockville, Md, Substance Abuse and Mental Health Services Administration. Available at nrepp.samhsa.gov/AboutNREPP.aspx. Accessed May 13, 2014Google Scholar

5 Green LW: From research to “best practices” in other settings and populations. American Journal of Health Behavior 25:165–178, 2001Crossref, MedlineGoogle Scholar

6 Bruns EJ, Hoagwood KE: State implementation of evidence-based practice for youths, part I. responses to the state of the evidence. Journal of the American Academy of Child and Adolescent Psychiatry 47:369–373, 2008Crossref, MedlineGoogle Scholar

7 Essock SM, Covell NH, Leckman-Westin E, et al.: Identifying clinically questionable psychotropic prescribing practices for Medicaid recipients in New York State. Psychiatric Services 60:1595–1602, 2009LinkGoogle Scholar

8 Aos S, Mayfield J, Miller M, et al.: Washington State Evidence-Based Treatment of Alcohol, Drug, and Mental Health Disorders: Potential Benefits, Costs and Fiscal Impacts on Washington State. Olympia, Washington State Institute for Public Policy, 2006Google Scholar

9 Ganju V: Implementation of evidence-based practices in state mental health systems: implications for research and effectiveness studies. Schizophrenia Bulletin 29:125–131, 2003Crossref, MedlineGoogle Scholar

10 Torrey WC, Drake RE, Dixon L, et al.: Implementing evidence-based practices for persons with severe mental illnesses. Psychiatric Services 52:45–50, 2001LinkGoogle Scholar

11 Hoagwood KE, Vogel JM, Levitt JM, et al.: Implementing an evidence-based trauma treatment in a state system after September 11: the CATS project. Journal of the American Academy of Child and Adolescent Psychiatry 46:773–779, 2007Crossref, MedlineGoogle Scholar

12 North MS, Gleacher AA, Radigan M, et al.: Evidence-Based Treatment Dissemination Center (EBTDC): bridging the research-practice gap in New York State. Report on Emotional and Behavioral Disorders in Youth 8:9–17, 2008Google Scholar

13 Gleacher AA, Nadeem E, Moy AJ, et al.: Statewide CBT training for clinicians and supervisors treating youth: the New York State Evidence Based Treatment Dissemination Center. Journal of Emotional and Behavioral Disorders 19:182–192, 2011Crossref, MedlineGoogle Scholar

14 OMH Clinic Reporting Participation—Internal Report Based on Analyses From SPARCS Data. Albany, New York State Office of Mental Health, 2010Google Scholar

15 McHugh RK, Barlow DH: The dissemination and implementation of evidence-based psychological treatments: a review of current efforts. American Psychologist 65:73–84, 2010Crossref, MedlineGoogle Scholar

16 Fixsen D, Blase K, Metz A, et al.: Statewide implementation of evidence-based programs. Exceptional Children 79:213–230, 2013CrossrefGoogle Scholar

17 Chor KHB, Olin SS, Weaver J, et al.: Adoption of clinical and business trainings by child mental health clinics in New York State. Psychiatric Services 65:1439–1444, 2014LinkGoogle Scholar

18 Wisdom JP, Chor KHB, Hoagwood KE, et al.: Innovation adoption: a review of theories and constructs. Administration and Policy in Mental Health 41:480–502, 2014Crossref, MedlineGoogle Scholar

19 Chor KHB, Wisdom JP, Olin SS et al.: Measures for predictors of innovation adoption. Administration and Policy in Mental Health (Epub ahead of print, April 17, 2014)Google Scholar

20 Aarons GA, Hurlburt M, Horwitz SM: Advancing a conceptual model of evidence-based practice implementation in public service sectors. Administration and Policy in Mental Health 38:4–23, 2011Crossref, MedlineGoogle Scholar

21 Panzano PC, Sweeney HA, Seffrin B, et al.: The assimilation of evidence-based healthcare innovations: a management-based perspective. Journal of Behavioral Health Services and Research 39:397–416, 2012Crossref, MedlineGoogle Scholar

22 Klein KJ, Sorra JS: The challenge of innovation implementation. Academy of Management Review 21:1055–1080, 1996CrossrefGoogle Scholar

23 Stata 11.2. College Station, Tex, StataCorp. Available at www.stata.com. Accessed March 13, 2014Google Scholar

24 Feldstein AC, Glasgow RE: A practical, robust implementation and sustainability model (PRISM) for integrating research findings into practice. Joint Commission Journal on Quality and Safety 34:228–243, 2008Google Scholar

25 Damschroder LJ, Aron DC, Keith RE, et al.: Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Science 4:50, 2009Crossref, MedlineGoogle Scholar