The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×
Published Online:

Psychosocial evidence-based practices can lead to desirable outcomes for people with serious mental illness. These outcomes include improved community and interpersonal functioning, better quality of life, and lessening of psychiatric symptoms ( 1 , 2 ). Unfortunately, these practices are not routinely available, and many people with serious mental illness are not receiving evidence-based care ( 3 , 4 ). Widespread variation at the state level has been reported with regard to commitment to and success of implementation of evidence-based practices ( 5 ). One hindrance to widespread dissemination of evidence-based practices is a lack of knowledge about the process of implementation ( 6 ).

The literature suggests that implementation of evidence-based practices has failed for numerous reasons. Most evidence-based practices are complex and may be difficult to implement without preexisting structure and support ( 7 ). Some studies have suggested that the principal reason for failure lies in an inadequate implementation plan with no clear model specification ( 8 , 9 ). A recent review of the mental health implementation literature found large conceptual and empirical gaps ( 10 ), highlighting the need for implementation research.

The National Implementing Evidence-Based Practices Project investigated the implementation of five psychosocial practices in routine mental health settings: supported employment ( 11 ), family psychoeducation ( 12 ), illness management and recovery ( 13 ), integrated dual disorders treatment ( 14 ), and assertive community treatment ( 15 ). The project examined practices implemented in 53 sites across eight states that used a common implementation model, which included material and human resources for each evidence-based practice. A mixed-methods study was conducted to understand the processes and outcomes of implementation over a two-year period at each site. Additional details about the project have been previously published ( 6 ).

The unit of analysis for the study reported here is the site, and the primary outcome is model fidelity ( 16 ). Fidelity scales have been validated for some evidence-based practices, such as assertive community treatment and supported employment, and have proven useful in differentiating among programs ( 17 ). Studies that used fidelity scales have found better outcomes for consumers when services adhere closely to a model with specified critical components and standards ( 18 , 19 , 20 , 21 , 22 , 23 , 24 , 25 ). This relationship has been established for assertive community treatment and supported employment but not for the other three practices included in this study.

This is the first prospective study with a sufficient number of sites to permit use of fidelity as the outcome of implementation in order to examine differences among evidence-based practices. Our primary aim in this article is to present the two-year fidelity results from the National Implementing Evidence-Based Practices Project. The study had two overarching purposes: to discern whether certain evidence-based practices were implemented more faithfully than others and to examine change over time in fidelity within each evidence-based practice in order to determine the critical time exposure for successful implementation.

Methods

Sites

Mental health authorities in eight states agreed to participate in the project, which involved recruiting sites and developing training and consultation capacity. Each state identified two evidence-based practices for dissemination. Various mechanisms were used across the states to recruit sites and to determine which sites would implement which practice. Some states chose among solicited proposals; others used less formal procedures. Sites were public-sector community mental health agencies, and the evidence-based practices were implemented within their programs of care for people with serious mental illness. The extent of practitioner involvement in and consumer access to the evidence-based practice was determined by each site and varied widely.

States provided a consultant-trainer for each evidence-based practice, and sites agreed in principle to provide time for training and supervision and to develop a relationship with the consultant-trainer. States gave no additional financial incentives to the sites, and they differed in the extent of their commitment and resources to disseminate and support evidence-based practices. Sites also agreed to participate in a range of evaluation activities, which were coordinated by an implementation monitor assigned to each site. Implementation monitors visited the sites monthly to collect systematic qualitative and quantitative data on the process and outcomes of implementation. Implementation monitors and consultant-trainers covered from two to eight sites within their states. Institutional review board approval was obtained for the overall project and within each state.

The baseline assessments and the start of implementation ranged from mid-2002 to mid-2003. Implementation monitoring continued for two years. Forty-nine sites provided two years of data.

Implementation model

The implementation model arose from a literature review, practical experience of services researchers, and focus groups with stakeholders. Additional details are available elsewhere ( 7 ). Central to the model were the implementation resource kits, also known as toolkits, and a consultant-trainer ( 26 ). The toolkits contained practice-specific and common resources, such as a user's guide and implementation tips for program leaders; introductory videos, PowerPoint presentations, and brochures; a practice demonstration video and workbook for practitioners; and fidelity scales with protocols.

The consultant-trainer provided training and clinical supervision to the program leaders and practitioners who were implementing the practice. The model prescribed a half-day kickoff session to introduce the practice to all stakeholders and three days of skills training modules for practitioners. Additional training and clinical supervision were provided as requested. The consultant-trainer also provided consultation to the leadership of the agency concerning concomitant organizational changes. Consultant-trainers were trained and supervised in both roles by evidence-based practice experts through monthly conference calls and semiannual in-person meetings.

The intensity and quality of the training and consultation varied both within and across states and practices, as did use of the toolkits. In addition to the kickoff and skills training, some sites received on-site training and consultation once per month for two years, whereas others received one day quarterly in the first year only. This variation was the result of individual differences among the consultant-trainers, the contexts in which they worked (states and sites), and the practices themselves.

Assessing fidelity

Fidelity was assessed by rating adherence to the principles and procedures specified in the evidence-based practice models. Fidelity scales had been validated previously for assertive community treatment ( 27 ) and supported employment ( 11 ). Investigators within this project, in conjunction with developers of the practices, created fidelity scales for integrated dual disorders treatment, illness management and recovery, and family psychoeducation. The implementation resource kits, including the fidelity scales, can be obtained on the Web site of the Center for Mental Health Services of the Substance Abuse and Mental Health Services Administration (mentalhealth.samhsa.gov/cmhs/communitysupport/toolkits).

The assessment of fidelity was similar across practices. It involved one-day site visits to gather information from various sources in order to make 5-point ratings on the critical components of the practice. A rating of 5 indicates full adherence to the model, and 1 indicates no adherence. The average of the item ratings yields a total fidelity score. For the study reported here, a total score of 4.0 or greater indicated high fidelity, scores between 3.0 and 4.0 indicated moderate fidelity, and scores less than 3.0 indicated low fidelity. Two trained raters—the consultant-trainer and the implementation monitor—assessed fidelity. Senior staff for each evidence-based practice provided initial training to the fidelity assessors and provided monthly telephone supervision.

Fidelity was assessed at baseline (before implementation) and at six, 12, 18, and 24 months thereafter. Assessors followed a detailed protocol with instructions for preparing sites for the visit, critical elements in the fidelity assessment, and sample interview questions. The protocol also included a fidelity assessment checklist. The assessment schedule typically included interviews with the team leader and practitioners, observation of team meetings and the intervention (for example, accompanying an assertive community treatment case manager), interviews with clients, and review of client charts. After the site visit, each assessor made independent fidelity ratings. The two assessors then reconciled any discrepancies to arrive at the final fidelity ratings. The consultant-trainer provided a fidelity report to the site after each assessment, which summarized the fidelity ratings and provided advice concerning components of the practice that were deficient.

Analysis

The interrater reliability of the fidelity scales was evaluated with the intraclass correlation coefficient (ICC) ( 28 ), based on a one-way random-effects analysis of variance (ANOVA) model for agreement between the two fidelity assessors on the total scale scores. The ICC was computed across all assessment points for each fidelity scale.

Forty-nine sites that completed the project were included in the outcome analyses. We first analyzed the fidelity outcomes by evidence-based practice at the end of the two-year implementation period. One-way ANOVA compared the average endpoint fidelity scores among the five evidence-based practices. Second, we examined change over time in fidelity for the five evidence-based practices. Mixed-effects regression models were used to test the significance of the time and practice main effects and their interaction.

Results

Table 1 shows the distribution of the eight states and 53 sites that set out to implement the five evidence-based practices.

Table 1 Number of states and sites for each of the evidence-based practices implemented in the National Implementing Evidence-Based Practices Project
Table 1 Number of states and sites for each of the evidence-based practices implemented in the National Implementing Evidence-Based Practices Project
Enlarge table

Interrater reliability

Interrater agreement between pairs of independent assessors was high for all five fidelity scales (assertive community treatment: ICC=.99, N=52 pairs of ratings; family psychoeducation: ICC=.99, N=24; integrated dual disorders treatment: ICC=.89, N=48; illness management and recovery: ICC=.97, N=50; and supported employment: ICC=.98, N=34). These results indicate high agreement between raters and minimal bias due to the relationship of the consultant-trainers with the sites.

Endpoint differences among evidence-based practice

Table 2 presents the summary statistics by evidence-based practice for the fidelity scales at the 24-month assessment point. Assertive community treatment, family psychoeducation, and supported employment were on average at or above the threshold value (4.0), and three-quarters or more of the sites attained high-fidelity implementation. Illness management and recovery and integrated dual disorders treatment fared less well overall. Their average fidelity scores were below the threshold, and their rates of high-fidelity implementation were below the other three practices. Overall, 29 of 49 sites (59%) scored 4.0 or higher on the respective fidelity scale at the two-year assessment point. If the four dropout sites are considered to have had low-fidelity implementation, the overall rate of high-fidelity implementation was 55% (29 of 53 sites). Specifically, the rate changes to 50% for family psychoeducation (three of six sites) and 15% for integrated dual disorders treatment (two of 13 sites).

Table 2 Scores on fidelity scales at 24 months and range of scores over two years for five evidence-based practices at 49 sites in the National Implementing Evidence-Based Practices Project
Table 2 Scores on fidelity scales at 24 months and range of scores over two years for five evidence-based practices at 49 sites in the National Implementing Evidence-Based Practices Project
Enlarge table

One-way ANOVA revealed significant differences at the endpoint among the five evidence-based practices in average fidelity scores (F=4.10, df=4 and 48, p<.007). Pairwise comparisons revealed few significant differences, but a significant difference was found when assertive community treatment and supported employment were compared with the other three practices. Despite mean differences among the practices at the endpoint, each one was implemented with high fidelity by some sites, as indicted by the number of sites with scores of 4.0 or higher and by the maxima in Table 2 . Moreover, only seven sites had endpoint fidelity below 3.0: none for assertive community treatment and family psychoeducation, one for supported employment, and three each for illness management and recovery and integrated dual disorders treatment.

Fidelity trends over time

Figure 1 shows longitudinal trends in average fidelity for each evidence-based practice over two years. Gains in fidelity were confined largely to year 1, although there were differences among the practices. Assertive community treatment leveled off after six months, supported employment and illness management and recovery leveled off after one year, and family psychoeducation and integrated dual disorders treatment had a positive slope throughout. Mixed-effects regression showed significant main effects for both linear time (F=127.74, df=1 and 44, p<.001) and quadratic time (F=72.80, df=1 and 44, p<.001) and for practice (F=12.46, df=4 and 44, p<.001), as well as significant interactions of practice with both linear time (F=4.02, df=4 and 44, p<.007) and quadratic time (F=2.90, df=4 and 44, p<.03).

Figure 1 Mean fidelity scores over two years for five evidence-based practices implemented at 49 sites

Because of the curvilinear nature of fidelity over time and to the interactions, separate mixed-effects regression models were fit to contrast the two years of implementation. An analysis including baseline, six, and 12 months (year 1) revealed a significant main effect for time (F=147.82, df=1 and 44, p<.001), whereas an analysis including 12, 18, and 24 months (year 2) did not. As Figure 1 shows, the evidence-based practices differed in the rate of change in fidelity during the first year, which produced a significant practice-by-time interaction (F=4.80, df=4 and 44, p<.003). The order of the evidence-based practices from the steepest linear slope to the shallowest was family psychoeducation, illness management and recovery, supported employment, assertive community treatment, and integrated dual disorders treatment. By year 2, four of the five evidence-based practices showed little further gain in fidelity. Only family psychoeducation improved substantially in year 2, but because of the small number of sites, the practice-by-time interaction was not significant. The main effect for practice was significant in each year, because assertive community treatment and supported employment had higher mean levels of fidelity throughout, but this effect was moderated by interactions with time.

Discussion

The 59% rate of high-fidelity implementation at two years is encouraging for the wider dissemination of these five practices. Only seven of the 49 sites (14%) failed to achieve an endpoint fidelity scale score of 3.0 or higher. Currently, there is no consensual agreement on cutoff scores for high fidelity, whether for quality assurance, implementation research, or accreditation. We chose a score of 4.0 or higher on face validity, but others have imposed different standards ( 17 , 29 , 30 ).

Because of the lack of comparison sites and because of confounding in this study (for example, states and consultant-trainers), the findings must be viewed as descriptive. They leave open questions concerning the effectiveness of the implementation model and differences among the five practices. Moreover, there is little relevant literature in the mental health field with which to compare these results. The rates of high-fidelity implementation compare favorably with those in previous studies—for example, 53% for assertive community treatment ( 17 ), 50% for family psychoeducation ( 31 ), and 61% for supported employment ( 32 ).

There are several possible reasons for the differences in endpoint fidelity. The practice models, and hence the fidelity scales, variously emphasize the structure of the practice versus the clinical expertise required to deliver it, which may account for baseline and endpoint differences. Supported employment and assertive community treatment require structural and clinical changes, whereas other practices rely more on clinical interventions. Structural changes can often be implemented quickly, whereas clinical skills require extensive training and supervision to implement fully.

Furthermore, the differences among the practices must be interpreted cautiously, because the fidelity scales have not been calibrated against each other. Item analysis and a larger sample of sites will be needed to fully evaluate the properties of the five fidelity scales. Despite this limitation, several factors support the credibility of the findings reported here: similar procedures were used to develop the fidelity scales, common assessment methods were used, and the scales have high face validity. Moreover, the fidelity scales provided the basis for implementation targets for the sites, and as the maxima in Table 2 indicate, high scores were attainable on all five scales.

The longitudinal fidelity results indicate that providers using the toolkit implementation model and similar resources should be able to achieve successful implementation within 12 months. The variation in the rate of implementation highlights another difference among the evidence-based practices. Unlike assertive community treatment and supported employment, which can attain high fidelity rapidly, illness management and recovery and family psychoeducation unfold more slowly over time in a prescribed fashion, and therefore it is impossible to achieve high fidelity in these practices until all of the stages have been implemented.

The toolkit model prescribed one year of training and consultation to achieve successful implementation, followed by one year of reduced support to facilitate sustaining of the practice. Accordingly, the longitudinal results revealed few further gains in fidelity during year 2, but they also revealed no erosion of the gains made during year 1. The focus at the sites shifted from improving fidelity to sustaining the practice, even if high-fidelity implementation had not been achieved by the end of year 1. This suggests the need to revise the implementation model to include booster training or other interventions at sites that fail to reach high fidelity in one year in order to facilitate further gains in fidelity and to support sustainability.

This study focused on model fidelity as the measure of evidence-based practice implementation, although there are other indicators of implementation success. The penetration of the practice provides another perspective on implementation, where penetration is defined as the proportion of eligible consumers who have access to the practice. Measures of consumer satisfaction with evidence-based practice services and practitioner attitudes toward evidence-based practices could also be used to evaluate implementation success. Another perspective on implementation concerns the quality of clinical skills and interactions, which are difficult to assess validly ( 33 ). System outcomes such as lower hospital and jail use should also accrue from successful implementation of evidence-based practices.

The ultimate aim of successful evidence-based practice implementation is to improve consumer outcomes, but outcomes are an imperfect measure of implementation, because the association between fidelity and outcomes is modest. Consequently, Goldman and colleagues ( 5 ) have cautioned that "measures of fidelity are a means to an end, not an end in themselves." Study of the relationship among multiple indicators of implementation is needed, in addition to study of the association between those indicators and consumer outcomes. An optimal approach to implementation that will improve consumer outcomes may be to combine approaches, that is, to seek fidelity to the model during early implementation and then to use outcome-based supervision to adapt and sustain the practice.

The limitations of this study must be considered when drawing inferences from the fidelity results. Sites were not selected randomly. All sites volunteered, but the rigor and nature of the selection process differed across states. No doubt those chosen were among the more motivated to implement a new practice, but they cannot be considered to be early adopters because all practices except illness management and recovery have been around for at least ten years. Nevertheless, the generalizability of the findings to other sites may be limited by selection factors. In addition, the fidelity assessors were not blinded. The consultant-trainers had an investment in the success of their sites, but the high agreement between them and the implementation monitors, who were not similarly invested, suggests the absence of strong and systematic bias.

In addition, the fidelity results are descriptive rather than analytic, and we cannot yet answer questions as to why certain evidence-based practices had higher fidelity implementation than others or why certain practices took longer on average to reach high fidelity. The differences may be due to the practices themselves, to contextual factors, or to the fidelity scales. In the course of this study, we accumulated a large amount of qualitative and quantitative data concerning the process of implementation at these sites. In follow-up papers, these data will be used to examine predictors of successful implementation and to explore the facilitators of, and barriers to, high-fidelity implementation both within and across evidence-based practices.

Conclusions

This is the first article to report a systematic study of the fidelity of evidence-based practice implementation across a large number of sites. Although we acknowledge limitations in our study and emphasize its descriptive nature, we have determined that most mental health centers can implement evidence-based practices and that one year appears to be the average amount of time needed for successful implementation to occur using the toolkit model of practice dissemination. Further, we found that some evidence-based practices may be easier to implement than others, but the validity of this conclusion awaits fidelity scale refinement, in-depth process analyses, and study replication.

Acknowledgments and disclosures

This research was supported by grant 036805 from the Robert Wood Johnson Foundation; contract 280-02-8070 from the Center for Mental Health Services, Substance Abuse and Mental Health Services Administration; a gift from the West Family Foundation; and the Johnson & Johnson-Dartmouth Community Mental Health Program. The authors thank the members of the National Implementing Evidence-Based Practices Project team and the participants at the sites in the eight states.

The authors report no competing interests.

Dr. McHugo, Dr. Drake, and Dr. Whitley are affiliated with the Departments of Psychiatry and Community and Family Medicine, Dartmouth Medical School. Dr. Bond and Ms. Campbell are with the Department of Psychology, Indiana University-Purdue University, Indianapolis. Dr. Rapp is with the School of Social Work, University of Kansas, Lawrence. Dr. Goldman is with the Department of Psychiatry, University of Maryland School of Medicine, Baltimore. Dr. Lutz is with the Ohio Department of Mental Health, Columbus. Dr. Finnerty is with the New York State Office of Mental Health, Albany. Send correspondence to Dr. McHugo at Dartmouth Psychiatric Research Center, 2 Whipple Pl., Ste. 202, Lebanon, NH 03766 (e-mail: [email protected]).

References

1. Drake RE, Goldman HH, Leff HS, et al: Implementing evidence-based practices in routine mental health service settings. Psychiatric Services 52:179–182, 2001Google Scholar

2. Lehman AF, Steinwachs DM, Co-Investigators of the PORT Project: Translating research into practice: the Schizophrenia Patient Outcomes Research Team (PORT) treatment recommendations. Schizophrenia Bulletin 24:1–10, 1998Google Scholar

3. Mental Health: A Report of the Surgeon General. Washington, DC, Department of Health and Human Services, US Public Health Service, 1999Google Scholar

4. Achieving the Promise: Transforming Mental Health Care in America. Pub no SMA-03-3832. Rockville, Md, Department of Health and Human Services, President's New Freedom Commission on Mental Health, 2003Google Scholar

5. Goldman HH, Ganju V, Drake RE, et al: Policy implications for implementing evidence-based practices. Psychiatric Services 52:1591–1597, 2001Google Scholar

6. Torrey WC, Lynde DW, Gorman P: Promoting the implementation of practices that are supported by research: the National Implementing Evidence-Based Practices Project. Child and Adolescent Clinics of North America 14:297–306, 2005Google Scholar

7. Torrey WC, Drake RE, Dixon L, et al: Implementing evidence-based practices for persons with severe mental illnesses. Psychiatric Services 52:45–50, 2001Google Scholar

8. Bond GR: Variations in an assertive outreach model. New Directions in Mental Health Services 52:65–80, 1991Google Scholar

9. Rosenheck R, Neale M, Leaf P, et al: Multisite experimental cost study of intensive psychiatric community care. Schizophrenia Bulletin 21:129–140, 1995Google Scholar

10. Fixsen DL, Naoon SF, Blase KA, et al: Implementation Research: A Synthesis of the Literature. FMHI pub no 231. Tampa, Fla, University of South Florida, Louis de la Parte Florida Mental Health Institute, National Implementation Research Network, 2005Google Scholar

11. Bond GR, Becker DR, Drake RE, et al: Implementing supported employment as an evidence-based practice. Psychiatric Services 52:313–322, 2001Google Scholar

12. Dixon L, McFarlane WR, Lefley H, et al: Evidence-based practices for services to families of people with psychiatric disabilities. Psychiatric Services 52:903–910, 2001Google Scholar

13. Mueser KT, Corrigan PW, Hilton DW, et al: Illness management and recovery: a review of the research. Psychiatric Services 53:1272–1284, 2002Google Scholar

14. Drake RE, Essock SM, Shaner A, et al: Implementing dual diagnosis services for clients with severe mental illness. Psychiatric Services 52:469–476, 2001Google Scholar

15. Phillips SD, Burns BJ, Edgar ER, et al: Moving assertive community treatment into standard practice. Psychiatric Services 52:771–779, 2001Google Scholar

16. Bond GR, Evans L, Salyers MP, et al: Measurement of fidelity in psychiatric rehabilitation. Mental Health Services Research 2:75–87, 2000Google Scholar

17. Salyers MP, Bond GR, Teague GB, et al: Is it ACT yet? Real-world examples of evaluating the degree of implementation for assertive community treatment. Journal of Behavioral Health Services and Research 30:304–320, 2003Google Scholar

18. Becker DR, Smith J, Tanzman B, et al: Fidelity of supported employment programs and employment outcomes. Psychiatric Services 52:834–836, 2001Google Scholar

19. Becker DR, Xie H, McHugo GJ, et al: What predicts supported employment program outcomes? Community Mental Health Journal 42:303–318, 2006Google Scholar

20. Bond GR, Salyers MP: Prediction of outcome from the Dartmouth ACT Fidelity Scale. CNS Spectrums 9:937–942, 2004Google Scholar

21. Henggeler SW, Pickrel SG, Brondino MJ: Multisystematic treatment of substance-abusing and-dependent delinquents: outcomes, treatment fidelity, and transportability. Mental Health Services Research 1:171–184, 1999Google Scholar

22. Jerrell JM, Ridgely MS: Impact of robustness of program implementation on outcomes of clients in dual diagnosis programs. Psychiatric Services 50:109–112, 1999Google Scholar

23. McHugo GJ, Drake RE, Teague GB, et al: Fidelity to assertive community treatment and client outcomes in the New Hampshire dual disorders study. Psychiatric Services 50:818–824, 1999Google Scholar

24. McDonnell J, Nofs D, Hardman M, et al: An analysis of the procedural components of supported employment programs associated with employment outcomes. Journal of Applied Behaviour Analysis 22:417–428, 1989Google Scholar

25. Oxman TE, Schulbert HC, Greenberg RL, et al: A fidelity measure for integrated management of depression in primary care. Medical Care 44:1030–1037, 2006Google Scholar

26. Mueser KT, Torrey WC, Lynde D, et al: Implementing evidence-based practices for people with severe mental illness. Behaviour Modification 27:387–411, 2003Google Scholar

27. Teague GB, Bond GR, Drake RE: Program fidelity in assertive community treatment: Development and use of a measure. American Journal of Orthopsychiatry 68:216–232, 1998Google Scholar

28. McGraw KO, Wong SP: Forming inferences about some intraclass correlation coefficients. Psychological Methods 1:30–43, 1996Google Scholar

29. Latimer E, Bush P, Becker DR, et al: How much does supported employment of the severely mentally ill cost? An exploratory survey of high-fidelity programs. Psychiatric Services 55:401–406, 2004Google Scholar

30. Rosenheck RA, Lam J, Morrissey JP, et al: Service systems integration and outcomes for mentally ill homeless persons in the ACCESS Program. Psychiatric Services 53:958–966, 2002Google Scholar

31. McFarlane WR, McNary S, Dixon L, et al: Predictors of dissemination of family psychoeducation in community mental health centers in Maine and Illinois. Psychiatric Services 52:935–942, 2001Google Scholar

32. Campbell K, Bond GR, Gervey R, et al: Does type of provider organization affect fidelity to evidence-based supported employment? Journal of Vocational Rehabilitation, in pressGoogle Scholar

33. Drake RE, Bond GR, Rapp CA: Explaining the variance within supported employment programs: comment on "What predicts supported employment outcomes?" Community Mental Health Journal 42:315–318, 2006Google Scholar