The de-implementation and persistence of low-value HIV prevention interventions in the United States: a cross-sectional study
Implementation Science Communications volume 1, Article number: 60 (2020)
As more effective or efficient interventions emerge out of scientific advancement to address a particular public health issue, it may be appropriate to de-implement low-value interventions, or interventions that are less effective or efficient. Furthermore, factors that contribute to appropriate de-implementation are not well identified. We examined the extent to which low-value interventions were de-implemented among public health organizations providing HIV prevention services, as well as explored socio-economic, organizational, and intervention characteristics associated with de-implementation.
We conducted an online cross-sectional survey from the fall of 2017 to the spring of 2019 with organizations (N = 188) providing HIV prevention services in the USA. Organizations were recruited from the Center for Disease Control and Prevention’s (CDC) website gettested.org from 20 metropolitan statistical areas with the highest HIV incidence. An organization was eligible to participate if the organization had provided at least one of the HIV prevention interventions identified as inefficient by the CDC in the last ten years, and one administrator familiar with HIV prevention programming at the organization was recruited to respond. Complete responses were analyzed to describe intervention de-implementation and identify organizational and intervention characteristics associated with de-implementation using logistic regression.
Organizations reported 359 instances of implementing low-value interventions. Out of the low-value interventions implemented, approximately 57% were group, 34% were individual, and 5% were community interventions. Of interventions implemented, 46% had been de-implemented. Although we examined a number of intervention and organizational factors thought to be associated with de-implementation, the only factor statistically associated with de-implementation was organization size, with larger organizations—those with 50+ FTEs—being 3.1 times more likely to de-implement than smaller organizations (95% CI 1.3–7.5).
While low-value interventions are frequently de-implemented among HIV prevention organizations, many persisted representing substantial inefficiency in HIV prevention service delivery. Further exploration is needed to understand why organizations may opt to continue low-value interventions and the factors that lead to de-implementation.
De-implementation is the process by which interventions and programs are discontinued and is an emerging line of inquiry within dissemination and implementation (D&I) research [1, 2]. Existing evidence suggests that de-implementation occurs in public health systems, but the extent to which de-implementation occurs when appropriate remains unclear [2,3,4]. Ideally, low-value interventions are de-implemented and replaced when either more effective interventions are available or more efficient, that is more cost-effective or streamlined, interventions are available, but well-integrated, low-value interventions persist in practice [1, 5]. Although there is considerable evidence that a multitude of factors at socio-economic, organizational, practitioner, and intervention levels contribute to the adoption and implementation of interventions , evidence outlining factors that make de-implementation more likely is limited in public health settings. Organizational capacity to provide services, for instance, funding, staff, and adequate collaborations with external partners are the most proximal factors contributing to service delivery . In addition, characteristics of interventions, such as interventions that are complicated or burdensome to provide may make de-implementation more or less likely .
HIV prevention provides an ideal public health context for exploring de-implementation of low-value interventions. Investment in HIV prevention research has led to several waves of evidence-based interventions over the last three decades [8,9,10,11]. The Centers for Disease Control and Prevention (CDC) has spearheaded HIV prevention efforts in the US by identifying the most effective and efficient interventions using rigorous effectiveness criteria, and then disseminating identified interventions . Historically, intervention has been disseminated by the CDC through two key policies. The first policy, the Diffusion of Effective Behavioral Interventions for HIV prevention (DEBI), beginning in 2002, supported an array of over 30 interventions aimed primarily at reducing sexual risk behaviors available for individual, group, and community levels of intervention as well as targeting different populations at increased risk for HIV in the US like intravenous drug users, specific ethnic and racial minorities, and men who have sex with men [12, 13], and more interventions were added in subsequent years. Drawing on the best available evidence at the time, uptake of interventions included as part of the DEBI policy was substantial and made evidence-based intervention ubiquitous in HIV prevention practice. These interventions have been disseminated to approximately 11,300 organizations including public health departments, community-based organizations, and medical clinics over the last 15 years . In supporting the DEBI policy, the CDC has invested approximately $100 million dollars identifying and disseminating HIV prevention interventions per year .
In 2012, the CDC revised HIV prevention intervention recommendations as part of the High Impact Prevention (HIP) policy. With emerging evidence supporting the effectiveness of biomedical interventions, especially pre-exposure prophylaxis, and evidence lacking for the effectiveness of some behavioral interventions, the HIP policy focused more heavily on expanding HIV testing and engaging/retaining HIV-positive individuals in clinical care . While some interventions recommended as part of the DEBI policy are still considered the most effective and were subsumed under HIP, approximately 37 interventions are considered low value, meaning they are not the most effective or efficient to provide and are no longer recommended . Although public health organizations were not overtly encouraged to de-implement low-value interventions no longer recommended, financial support (a key factor leading to sustainment and de-implementation) was redirected toward HIP interventions [7, 16]. However, it is unclear if low-value HIV prevention interventions persist in public health organizations.
With the goal of quantitatively characterizing the extent of de-implementation through a broad survey of public health organizations involved in HIV prevention, we examined the extent to which low-value interventions were de-implemented among public health organizations providing HIV prevention services, as well as explored socio-economic, organizational, and intervention characteristics associated with de-implementation.
We report data from the quantitative first stage of a sequential mixed-methods study. We conducted a cross-sectional survey with 188 organizations through Qualtrics . Organizations were identified through the CDC’s website gettested.cdc.gov, a searchable website where organizations can register as providing an HIV test, a core HIV prevention and treatment service . We searched for organizations within a 50-mile radius, the widest search radius available, from the 20 metropolitan statistical areas (MSAs) with the highest HIV incidence. These 20 areas encompass approximately 50% of all new HIV diagnoses in the US . All research protocols were reviewed and approved by the Institutional Review Board at Washington University in St. Louis. When required, we also received approval from individual states to conduct research with local health departments. We used the STROBE and CHERRIES checklists for cross-sectional and online studies to develop and report our results in the current manuscript [20, 21].
Recruitment and eligibility
All organizations were recruited for participation continuously over the data collection period. Organizations were eligible if one of 37 low-value interventions defined as low value by the CDC had been implemented. Interventions included as part of the eligibility criteria are listed in Appendix A. Those reported as implemented are also shown in Fig. 1. Executive directors or supervisors who oversee HIV prevention services were recruited to complete the survey either by phone or email. If organizations had multiple locations, one person from the organization was asked to respond on behalf of all locations. Participants agreeing to participate were given a specific survey link. The consent document was provided at the start of the survey and in the recruitment email describing the purpose of the survey, the principal investigators, the length of the study, time allowed to respond, and details of data storage and reporting. Participants were asked an initial set of screening questions as part of the survey to determine (1) if the agency implemented a low-value intervention within the last ten years, and (2) if the contact was the most appropriate person to complete the full survey. If the recruited person indicated that they were not the most appropriate person to respond to the survey, we contacted the participant again and asked to be referred to the most appropriate staff member. Staff members were contacted a maximum of three times before considered a non-respondent.
Survey design and data collection
All data were self-reported via a survey conducted online through Qualtrics. Items for the survey were generated from existing questions  and existing implementation frameworks. Three frameworks were used: Meyer, Davis, and Mays’ conceptual framework for Organizational Capacity , Scheirer and Dearing’s framework for Sustainability of Public Health Programs , and Fixsen and colleagues’ implementation framework [4, 23]. Synthesized together, these frameworks identify the predictors of de-implementation within organizations, de-implementation of interventions as an outcome, and the downstream influences of de-implementation on communities and organizations as well as the processes involved in the de-implementation. Meyer and colleagues' Organizational Capacity Framework identifies central concepts (e.g., financial, human, and physical resources) indicated by specific variables essential to implementing and sustaining interventions over time. For example, financial resources are indicated by both level and source (external vs. internal). A reduction in capacity in one or more areas may lead an organization to de-implement an intervention . Our previous work and the Sustainability framework (the converse of de-implementation) suggests that when EBIs are de-implemented, there are subsequent outcomes for organizations, staff, clients and communities [1, 22]. The Implementation Framework captures the detailed processes involved with implementing and sustaining interventions in organizations . We extended this model to incorporate a de-implementation phase that describes the process of de-implementation and a set of activities likely to take place when de-implementation is occurring . Survey questions are available in Appendix B.
The survey protocol was piloted in a separate set of locales with approximately 20 executive directors and supervisors of HIV prevention services to identify any problems with the survey structure and flow. The survey question format was divided into several subsections: characteristics of the organization (e.g., type, size, client population) and of the respondent (e.g., age, gender, and race/ethnicity) (15 items), details of initial intervention implementation, and whether the intervention was continued or de-implemented (2 items). Then, the respondent was asked a series of questions for each intervention reported as continued (3 items) and each intervention reported as de-implemented (10 items). Depending on the number of interventions on which the participant was reporting, the total number of items ranged from 30 to 150 items. Surveys took on average 20 min to complete, ranging 15–30 min. To reduce reporting bias, participants were made aware that their responses would be kept confidential from their organization and when reporting study results. Participants were given a $20 gift card as incentive for completing the survey. Participants were able to change their answers through a back button. Participants were reminded up to three times to complete the survey, and the survey was closed if not completed after three weeks from the final reminder. Data were collected from the fall of 2017 to the spring of 2019.
Our primary goal was to describe the continuation or de-implementation of low-value interventions that had been adopted by HIV services organizations over the previous 10-year period. We looked only at the 37 interventions deemed low-value by the CDC, and examined the likelihood of intervention de-implementation based on organizational characteristics specifically organizational type, size, collaboration partners, funding sources, and type of intervention. Respondents chose their organization type in response to the question, “How would you best describe your organization?” with available choices of a community-based organization (CBO), Clinic, public health department (PHD), or other. Ten organizations chose “Other” and these were recoded into the appropriate categories. Most were either “non-profit” or some type of “hospital”, reclassified as CBO or Clinic. As an indication of organizational size, we used full-time equivalent (FTE). FTE was reported as a continuous variable and was collapsed into a range of categories (0–10, 11–50, 50 or more). We chose to collapse this variable as half of the organizations had fewer than 50 FTEs, three-quarters had fewer than 100, and two organizations were disproportionately large (10,000 and 34,000 FTEs). Our primary concern was whether small or large organizations would be more likely to de-implement interventions, rather than the specific number of FTEs. Respondents were asked if collaborations were either formal or informal with a number of other organizational types (a college or university, other health service agency, social service agency, faith-based organization, for-profit business, local or state health department, or other) and funding sources (federal government, state government, non-profit, community/individual donations, insurance billing, direct fees, or other). An intervention was categorized as individual-level if the interventions was intended to be delivered one-on-one with a client, a group-level if the intervention was intended to be delivered with more than one client, and a community-level if the intervention was intended to be delivered with a number of people outside of the organization.
We downloaded data from Qualtrics and used the R statistical environment for analysis . Complete responses were used for analyses. We provide descriptive statistics indicating the frequency, type of interventions implemented, and whether they were continued or de-implemented. Organizational and intervention characteristics and their association with de-implementation were investigated via logistic regression.
Organizations and participants
We identified 876 HIV service providers within 50 miles of the 20 metropolitan areas, recruited all of them via phone or email or both, and achieved a response rate of 66% (N = 578), see Fig. 1. Of organizations that responded, 38% (N = 213) met the eligibility criteria, meaning they implemented one of the low-value interventions we identified. Five percent (N = 51) of organizations declined to participate. Of the 213 eligible respondents, 12% (N = 25) did not finish the survey, leaving 188 eligible responses available for analysis. Respondents representing organizations were diverse in race and ethnicity (25% black, 34% white, 22% Hispanic, and 8% other), education level (15% < 4-year degree, 27% 4-year degree, 59% graduate degree), and age (38% 20–39 years, 47% 40–59 years, 15% 60+ years). Participants were 62% female and reported an average of 11.1 years with their current agency.
Table 1 shows the interventions reported by intervention level, organization type, and organization size. A slight majority of organizations were community-based (52.7%), while 29.8% were clinics, including hospitals and federally qualified healthcare centers, and the remaining 17.6% were state or local departments of health. About one fifth (21%) of the organizations had fewer than 10 FTEs, 30% had 10 to 50 FTEs, and the remaining 41% had more than 50 FTEs. Most organizations had active collaborations of one kind or another, most frequently social (80.9%) or health service (88.8%) organizations, and 63.3% of organizations had collaborations with a college or university. Most reported a variety of funding sources, with state or local (88.3%) and federal (73.9%) funds being most often reported. While there was much variation in the percentage of priority populations served, low socio-economic status and racial or ethnic minority individuals on average comprised the largest portions of clientele.
Low-value interventions implemented
Over the previous 10 years, organizations adopted a variety of interventions initially recommended but subsequently deemed low value. In all, the 188 organizations reported initially implementing 359 low-value interventions, most of which were group-level (56.8%), and the smallest number 16 (4.5%) were community-level interventions (Table 2). About half of the organizations (51%; N = 95) offered only one of the interventions, and the remaining 93 offered two or more. The average number of interventions implemented per organization was 1.91 (SD = 1.4), the median number of interventions was 1.
Figure 2 shows the number of times each specific intervention was adopted and implemented by the responding organizations. The most frequently reported intervention was Behavioral Risk Counseling, an individual-level intervention reported by 128 agencies (68%); however, a variety of group-level interventions were implemented (22 different interventions). Of these, SISTA, an intervention designed for black women, and Safer Sex, an intervention designed for youth, were reported most often, by 47 (25%) and 38 (38%) organizations, respectively.
Figure 3 shows the current implementation status—de-implemented or continued—of reported interventions overall and by characteristics of the organizations. The first row of the bar chart shows that overall almost half of the 359 low-value interventions were de-implemented (46%) indicated by the dark bars and 54% continued. By intervention type, 39% of individual interventions were de-implemented, while half (50% or 102) of group-level and 10 of the 16 (62%) reported community-level interventions were de-implemented. Below that examining federal funding support and academic affiliation, we observed very little variation in intervention de-implementation between organizations that received federal funding or had an academic affiliation and those that did not. The fourth row of charts examining organizational size shows that smaller organizations, those with 1–10 or 10–50 FTEs, de-implemented at smaller rates than their larger counterparts: 38% of interventions were de-implemented in organizations with 10 or fewer FTEs. In the bottom row of Fig. 3, we see that CBOs had the highest rates of de-implementation (50%) and departments of public health had the lowest rates (38%). Clinics, with similar numbers of group and individual interventions, had almost double the rate of de-implementation for group-level programs (55%) than that for individual ones (29%).
To examine the impact of these factors on de-implementation, we conducted a logistic regression (Table 3).
The outcome for “success” was de-implementation of a low-value intervention. We included organizational characteristics—organization type, size in FTEs, and binary variables for whether federal US funding was received and the existence of an affiliation with a college or university—along with the intervention level and the total number of interventions reported by each organization. Standard errors are clustered within organizations to account for any association among the interventions offered at a particular organization and their likelihood of de-implementation. Of the 93 organizations that offered two or more interventions, 59 organizations reported all of those offered had the same status (either all continued or all de-implemented) and the remaining 34 organizations reported different statuses for interventions. From the table, we see that all else equal, organization type, whether the organization was federally funded or affiliated with an educational institution, the level of the intervention, or the number of interventions offered was not associated with the probability of de-implementation at a 95% confidence level. However, organization size was related to the likelihood of ending programs. Controlling for the other factors, organizations with 50 or more FTEs were about three times as likely as small organizations (those with 10 or fewer FTEs) to de-implement programs.
Understanding the de-implementation of low-value interventions is essential as we continue to invest in basic research and promote the use of effective interventions in public health practice. There are still only a handful of studies examining de-implementation in the context of public health , and we extend this work by describing the extent to which de-implementation occurs for a public health issue with a wide array of low-value interventions (i.e., HIV) and describe factors that may (or may not) be associated with de-implementation.
Our results demonstrate that a range of organizations serving a variety of populations implemented a number of interventions, which were the best available interventions at the time. Clinics and CBOs were more likely to have implemented these interventions, likely because large municipal public health departments often give small grants to provide services rather than provide services directly. Organizations opted to continue these low-value interventions, now no longer considered the most effective or efficient, approximately half of the time. This rate of low-value intervention persistence represents substantial inefficient use of an organization’s capacity to provide services both in terms of staff time and money spent. We found this rate surprising since evidence against these interventions and subsequent loss of CDC support for these interventions occurred beginning in approximately 2014 .
Understanding the potential drivers of de-implementation in addition to the scientific evidence is an important precursor to successfully stimulating appropriate de-implementation . Toward this end, we examined whether specific organizational characteristics predicted appropriate de-implementation; however, very few organizational characteristics were associated with de-implementing low-value interventions. Organizational size was the only significant predictor of appropriate de-implementation with larger organizations (50 or more FTEs) more likely to de-implement low-value interventions. Larger organizations may have greater general capacity to make service delivery changes more quickly or have larger information networks, which leads to faster dissemination of information and evidence [7, 26].
Of equal interest is the number of organizational factors that did not demonstrate an association with de-implementation, namely academic partnerships or receipt of federal US funding. Ideally, academic partners and national agencies would support the dissemination of the most current evidence for or against any given intervention and incentivize change through funding mechanisms and technical support. This suggests external dissemination efforts to raise awareness about low-value interventions and availability of more efficient or effective interventions is inadequate or insufficient in promoting change similar to adoption of interventions. Furthermore, withdrawal of support by a national-level government agency, like the CDC, even after several years for specific interventions, is also likely insufficient to promote de-implementation among a substantial number of organizations. Organizations like clinics and CBOs, which have multiple funding streams, may opt to continue supporting an intervention through other means if one particular source of funding ends. Lastly, while descriptive statistics initially suggested that individual-level interventions were more likely to be continued and group- and community-level interventions de-implemented, this influence disappeared after controlling for organizational factors, namely organizational size. However, we do not consider these results definitive and would encourage others to continue exploring these factors to either confirm or refute our results in different public health settings with different types of interventions. Qualitative methods in particular with rich data collection could potentially help elucidate and explain relationships observed in this study.
Our study highlights a need to develop strategies that actively encourage appropriate de-implementation of ineffective or inefficient interventions. Healthcare settings, where research focusing on de-implementation and reduction of excess healthcare interventions has been ongoing for some time may provide promising approaches to de-implementation that could be tested in public . As others have noted, the process for selecting which interventions to provide or end is nuanced, and administrators and staff may be considering multiple factors given the perceived needs of the community, the multitude of intervention possibilities, and the capacity of the organization to make service delivery changes . Understanding from a staff perspective the reasons for continuing or de-implementing interventions and how staff make decisions to choose certain services over others is critical. Furthermore, understanding from a staff perspective why these interventions may continue to hold value will be helpful for public health scientists to develop useful strategies for encouraging de-implementation and support public health practitioners in their effort to provide the most effective services in the most efficient manner possible .
There were several limitations to the study. The study was not a random sample of all organizations delivering HIV services, which reduces the generalizability of the results. However, our wide coverage of areas nationally with high HIV incidence and diverse set of organizations helps ensure that we captured an appropriate cross section of organizations involved in HIV prevention. The study was conducted in a single country which limits the generalizability to other countries. Responses were self-report, which may have led to some recall bias. Also, while funding was often cited as a reason to end or continue interventions, we did not have data on funding amounts by source, which would have added to our analyses and findings.
In the emerging field of de-implementation in public health, we demonstrated that while appropriate de-implementation of low-value interventions occurs, intervention persistence also occurs at a high rate. We also demonstrated that few organizational characteristics predict appropriate intervention de-implementation. Identification of other factors driving service delivery decisions is needed to improve the appropriate de-implementation of low-value interventions, enhance the uptake of more effective programs, and ensure more cost-effective investment in public health.
Availability of data and materials
Data is available upon reasonable request to the first author.
Centers for Disease Control and Prevention
Diffusion of Effective Behavioral Interventions
Dissemination and implementation
High Impact Prevention
Human immunodeficiency virus
Public health department
Metropolitan statistical area
McKay VR, et al. Letting go: conceptualizing intervention de-implementation in public health and social service settings. Am J Community Psychol. 2018;62(1-2):189–202.
Niven DJ, et al. Towards understanding the de-adoption of low-value clinical practices: a scoping review. BMC Med. 2015;13(1):1–21.
Brownson RC, et al. Understanding mis-implementation in public health practice. Am J Prev Med. 2015;48(5):543–51.
McKay, V.R., M. Margaret Dolcini, and L.D. Hoffer, The dynamics of de-adoption: a case study of policy change, de-adoption, and replacement of an evidence-based HIV intervention. Transl Behav Med, 2017: p. 1-11.
Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. 2013;8(1):117.
Damschroder LJ, et al. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4.
Meyer A-M, Davis M, Mays GP. Defining organizational capacity for public health services and systems research. J Public Health Manag Pract. 2012;18(6):535–44.
Wingood GM, DiClemente RJ. Enhancing adoption of evidence-based HIV interventions: promotion of a suite of HIV prevention interventions for African American women. AIDS Educ Prev. 2006;18(4 Suppl A):161–70.
Sogolow, et al., Strengthening HIV prevention: application of a research-to-practice framework. 2000, New York, NY, ETATS-UNIS: Guilford. 149.
Collins CB, Sapiano TN. Lessons learned from dissemination of evidence-based interventions for HIV prevention. Am J Prev Med. 2016;51(4):S140–7.
Purcell DW, McCray E, Mermin J. The shift to high-impact HIV prevention by health departments in the United States. Public Health Rep. 2016;131(1):7–10.
McKleroy VS, et al. Adapting evidence-based behavioral interventions for new settings and target populations. AIDS Educ Prev. 2006;18:59–73.
Collins C, et al. The diffusion of effective behavioral interventions project: development, implementation, and lessons learned. AIDS Educ Prev. 2006;18(4 Suppl A):5–20.
Johns DM, Bayer R, Fairchild AL. Evidence and the politics of deimplementation: the rise and decline of the “counseling and testing” paradigm for HIV prevention at the US Centers for Disease Control and Prevention. Milbank Q. 2016;94(1):126–62.
CDC. High impact HIV prevention. 2015 [cited 2016 4/11/2016]; Available from: https://effectiveinterventions.cdc.gov/docs/default-source/general-docs/15-1106-hip-overview-factsheet.pdf?sfvrsn=2.
Flaspohler P, et al. Unpacking prevention capacity: an intersection of research-to-practice models and community-centered models. Am J Community Psychol. 2008;41(3-4):182–96.
Qualtrics. Qualtrics: Welcome to the experience management platform. 2017 [cited 2017 5/1/2017]; Available from: https://www.qualtrics.com/homepage/.
CDC. GetTested: national HIV, STD, and hepatitis testing. . N.D. [cited 2017; Available from: https://gettested.cdc.gov/search_results.
CDC. HIV surveillance report, 2015; vol. 27 2016 [cited 2017 5/1/2017]; Available from: http://www.cdc.gov/hiv/library/reports/hiv-surveillance.html. .
von Elm E, et al. The strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. Ann Intern Med. 2007;147(8):573–7.
Eysenbach G. Improving the quality of web surveys: the checklist for reporting results of internet E-surveys (CHERRIES). J Med Internet Res. 2004;6(3):e34.
Scheirer MA, Dearing JW. An agenda for research on the sustainability of public health programs. Am J Public Health. 2011;101(11):2059–67.
Bertram RM, Blase KA, Fixsen DL. Improving programs and outcomes: implementation frameworks and organization change. Res Soc Work Pract. 2015;25(4):477–87.
Team, R.C. R: a language and environment for statistical computing. R Foundation for statistical computing,. 2018; Available from: https://www.R-project.org/.
Montini T, Graham ID. “Entrenched practices and other biases”: unpacking the historical, economic, professional, and social resistance to de-implementation. Implement Sci. 2015;10.
Rogers EM, et al. Complex adaptive systems and the diffusion of innovations. Innov J. 2005;10(3):1–26.
Colla CH, et al. Interventions aimed at reducing use of low-value health services: a systematic review. Med Care Res Rev. 2017;74(5):507–50.
Pinto RM, Witte SS. No easy answers: avoiding potential pitfalls of de-implementation. Am J Community Psychol. 2019;63(1-2):239–42.
We would like to thank Enola Proctor for providing her mentorship in the conception of this project and Bradley Stoner for his oversight during the project.
This work has been supported through the National Institute of Mental Health award number R21MH115772-01; National Institute of Diabetes and Digestive and Kidney Diseases award numbers 5R01DK109913, 2P30DK092949, and P30DK092950; National Cancer Institute award number R01CA214530.
Ethics approval and consent to participate
All research protocols were reviewed and approved by the Institutional Review Board at Washington University in St. Louis. When required, we also received approval from individual states to conduct research with local health departments.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
McKay, V.R., Combs, T.B., Dolcini, M.M. et al. The de-implementation and persistence of low-value HIV prevention interventions in the United States: a cross-sectional study. Implement Sci Commun 1, 60 (2020). https://doi.org/10.1186/s43058-020-00040-6