The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×
Open ForumFull Access

Stop Calling Them Laggards: Strategies for Encouraging Nonadopters to Incorporate Evidence-Based Practices

Published Online:https://doi.org/10.1176/appi.ps.201900031

Abstract

Diffusion theory has deeply influenced implementation science and, specifically, language and thinking around practitioners and organizations that do not adopt evidence-based practices. In this Open Forum, the authors address the use of the term “laggard” to characterize this late or nonadopting group and argue for new nomenclature and modernized attitudes and approaches toward this group.

Few evidence-based practices (EBPs) reach widespread use in clinical settings. The estimates for proliferation of evidence-based psychosocial treatments in public health settings is particularly bleak. In recognition of the need for research to more directly affect clinical care, implementation science has developed frameworks and methods to support the systematic uptake of EBPs into routine practice (1). A classic theory that has deeply influenced the core tenets and language of implementation science is diffusion of innovations. Everett Rogers (2) developed diffusion of innovations theory more than 50 years ago to explain what factors impede or facilitate uptake of innovations, which Rogers refers to as adoption. He classifies individuals based on the speed with which they adopt new ideas or technologies. The five adopter categories are innovators, early adopters, early majority, late majority, and laggards. Laggards, also referred to as late or nonadopters, are the last in a social system to adopt an innovation. Rogers describes laggards as “possessing no opinion leadership,” “isolates,” “resistant,” and “uneducated.” Late and nonadopters of EBPs are understudied (3) and, we argue, largely misrepresented by implementation scientists and health services researchers to their own detriment. Diffusion of innovations is a critically important theory that has deeply influenced implementation science. We take issue with the term “laggard,” however, and wonder if the legacy of diffusion of innovations has resulted in a mischaracterization of this nonadopting group.

The Facts About Nonadopters

We recently published a study of non- and late-adopting providers in community mental health (4), and we have completed data collection on another study of nonadopters of pharmacotherapy for opioid use disorder. We find these individuals can be characterized by nearly none of the adjectives Rogers used. First, they are not uneducated. In many cases, they are aware of the scientific evidence for a particular therapy and they refute it based on their own experience and the experience of others in their group. Neither are they isolated, nor do they lack opinion leadership (particularly in their provider network). If they are resistant, it is to researchers telling them how to treat their clients or to the idea of the superiority of internally valid empirical evidence over their own experience. Fidelity to a psychosocial practice is not seen as relevant, meaningful, obtainable, or of value to patient outcomes. Although we are troubled that more systematically collected, rigorous evidence was not the main driver of clinical care in these organizations, we felt empathy for these so-called laggards because we think they are unfairly portrayed and have valuable insights to offer. We believe that the field’s failure to acknowledge their viewpoint has hampered our efforts toward a more representative and thorough approach to implementation research.

There are three reasons why we think this adopter group deserves further examination, reconceptualization, and certainly a new moniker. First, Rogers described laggards as the tail end of a continuum of adoption and adopter categories, representing 16% of the population. Among mental health and substance abuse treatment providers, however, laggards (defined as providers who do not adopt EBPs) represent most community providers (5). In other words, they are not lagging behind, they are the mode. Second, the terms “laggard” and “late adopter” imply that they will ultimately adopt. In reality, they are not simply “late” compared with other adopting groups; they may have made a conscious decision to not adopt under current circumstances. The term “laggard” perpetuates the top-down pipeline concept of disseminating research to practice and that providers are “empty vessels” who will eventually come around to the value of EBPs (6). This characterization exacerbates the misperception that nonadopting providers are ignorant of the evidence base and leads to the development of ineffective implementation strategies such as repeated presentations of statistical evidence, or policy mandates that require use of EBPs. Third, and perhaps most important, by failing to liaise with this important nonadopting group, we miss the benefits that come from incorporating different practice viewpoints to inform and educate us about the design of implementation strategies to enhance uptake of the treatments about which we feel so passionately.

Eliciting and incorporating providers’ input through dynamic partnership to address the fundamental disconnect between researcher and provider is foundational to implementation science and community-partnered research (6, 7). Efforts have included active strategies to engage frontline providers and to account for the inevitable modifications and adaptations of EBPs in real-world settings (8, 9). Providers who participate in most EBP implementation studies, initiatives, and partnerships, however, often share a belief that EBPs comprise an important component of maximizing outcomes for people with psychiatric symptoms and that the use of EBPs should be increased. Our experience suggests that many nonadopters do not share this understanding. In many cases, this provider group does not value EBPs (at least the way that researchers conceptualize them) and has no interest in implementing them. Community-partnered research offers important principles and models for connecting with nonadopters. Researchers who wish to develop implementation strategies aimed at nonadopters need to approach them cognizant that there may be little shared understanding of the value of EBPs and that these providers may be actively disinterested in collaborating with a group whose ultimate goal is to change the provider’s practice. Consistent with the first principles of community-partnered research, these researchers must spend considerable time learning about the values of nonadopters and take very different approaches to implementation than traditionally have been tested.

The results of our nonadopter studies indicate that nonadopters are much more educated and aware than they often are depicted to be and highlight a need for designing and selecting implementation strategies that target the causal mechanisms and unique driving forces that inhibit EBP use in this group. Traditional implementation efforts include strategies like developing education materials, training, academic detailing, auditing and giving feedback, implementing learning collaboratives, and improving implementation leadership (10). These are important strategies when targeted appropriately. It is highly unlikely, however, that these are effective avenues to secure buy-in or behavior change among late or nonadopters who refute the scientific evidence or see it as not applicable to the population they serve. Indeed, late adopters who were recipients of these strategies told us in our studies that they found data presentations on EBPs to be “remedial,” “heavy handed,” and “biased” and that they only served to increase their resistance. More research is sorely needed on the deployment of implementation strategies that affect attitudes toward traditional research findings and address and embrace ideology by broadening the latitude of acceptance toward research evidence.

Working With Nonadopters

We need to rigorously explore persuasion techniques that increase researchers’ credibility among nonadopters. Vivid case presentations, for example, have shown some promise in this regard (11). EBPs may be more palatable if they are designed as tools to be integrated into the art and ideology of helping the suffering. The impact of patient testimonials, peer influence, and consumer demand for EBPs on provider attitudes is still unknown and virtually unexplored. We need to listen carefully to the viewpoint of nonadopters, be responsive to (rather than dismiss) barriers they raise, and work with them to share an understanding of the problem we are trying to fix. Almost all of our nonadopters report that what they do works because they see it work every day. We need to explore strategies that show providers their own data (and those of other local providers who see similar types of patients). If what they are doing is effective, we should be partnering with them to investigate, evaluate, and replicate the benefits of their treatment innovations. If what nonadopters are doing is not effective, we need to be able to prove it by using their own data and develop effective strategies to encourage them to believe the data we share. There is room in the field to develop and test new approaches to practice-based research.

Conclusions

Nonadopters have much to offer in teaching researchers about attitudes toward and the viability of EBPs in their settings. Joint collaborations with nonadopters can inform implementation strategies across the adopter continuum. It is imperative, therefore, that we include their heretofore underrepresented perspective in the design, selection, and implementation of any strategy to enhance the uptake of research in any setting. It is high time to retire the term “laggard” and bring this group into the conversation about implementation, both at the level of science and at the policy table. By mischaracterizing those who do not adopt EBPs as ignorant or isolated, we hurt them and we hurt ourselves in scholarship, but most of all, we hurt the people who suffer from behavioral health disorders. At the very least, let’s stop calling them laggards.

Department of Psychiatry, Perelman School of Medicine, and Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia.
Send correspondence to Dr. Stewart ().

The study was funded by the National Institutes of Health (K23MH099179 and K23DA048167).

Dr. Beidas has received consulting fees from Merck, Sharpe, and Dohme and the Camden Coalition of Healthcare Providers and royalties from Oxford University Press. The other authors report no financial relationships with commercial interests.

References

1 Eccles MP, Mittman BS: Welcome to implementation science. Implement Sci 2006; 1:1CrossrefGoogle Scholar

2 Rogers E: Diffusion of Innovations, 5th ed. New York, Simon & Schuster, 2003Google Scholar

3 Aarons GA: Measuring provider attitudes toward evidence-based practice: consideration of organizational context and individual differences. Child Adolesc Psychiatr Clin N Am 2005; 14:255–271Crossref, MedlineGoogle Scholar

4 Stewart RE, Adams DR, Mandell DS, et al: Non-participants in policy efforts to promote evidence-based practices in a large behavioral health system. Implement Sci 2017; 12:70Google Scholar

5 Knudsen HK, Abraham AJ, Oser CB: Barriers to the implementation of medication-assisted treatment for substance use disorders: the importance of funding policies and medical infrastructure. Eval Program Plann 2011; 34:375–381Crossref, MedlineGoogle Scholar

6 Green LW: Making research relevant: if it is an evidence-based practice, where’s the practice-based evidence? Fam Pract 2008; 25(suppl 1):i20–i24Crossref, MedlineGoogle Scholar

7 Pellecchia M, Mandell DS, Nuske HJ, et al.: Community-academic partnerships in implementation research. J Community Psychol 2018; 46:941–952Google Scholar

8 Lyon AR, Koerner K: User-centered design for psychosocial intervention development and implementation. Clin Psychol 2016; 23:180–200Google Scholar

9 Stirman SW, Miller CJ, Toder K, et al.: Development of a framework and coding system for modifications and adaptations of evidence-based interventions. Implement Sci 2013; 8:65Crossref, MedlineGoogle Scholar

10 Powell BJ, Waltz TJ, Chinman MJ, et al.: A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci 2015; 10:21Crossref, MedlineGoogle Scholar

11 Stewart RE, Chambless DL: Interesting practitioners in training in empirically supported treatments: research reviews versus case studies. J Clin Psychol 2010; 66:73–95MedlineGoogle Scholar