- Open Access
Using decision analysis to support implementation planning in research and practice
Implementation Science Communications volume 3, Article number: 83 (2022)
The process of implementing evidence-based interventions, programs, and policies is difficult and complex. Planning for implementation is critical and likely plays a key role in the long-term impact and sustainability of interventions in practice. However, implementation planning is also difficult. Implementors must choose what to implement and how best to implement it, and each choice has costs and consequences to consider. As a step towards supporting structured and organized implementation planning, we advocate for increased use of decision analysis.
When applied to implementation planning, decision analysis guides users to explicitly define the problem of interest, outline different plans (e.g., interventions/actions, implementation strategies, timelines), and assess the potential outcomes under each alternative in their context. We ground our discussion of decision analysis in the PROACTIVE framework, which guides teams through key steps in decision analyses. This framework includes three phases: (1) definition of the decision problems and overall objectives with purposeful stakeholder engagement, (2) identification and comparison of different alternatives, and (3) synthesis of information on each alternative, incorporating uncertainty. We present three examples to illustrate the breadth of relevant decision analysis approaches to implementation planning.
To further the use of decision analysis for implementation planning, we suggest areas for future research and practice: embrace model thinking; build the business case for decision analysis; identify when, how, and for whom decision analysis is more or less useful; improve reporting and transparency of cost data; and increase collaborative opportunities and training.
Early implementation involves many choices [1,2,3]. These choices involve questions such as what intervention or evidence-based program (EBP) should be pursued for a given health issue of interest, or, if the intervention is already selected, what implementation strategies will best support success. Combined intervention/implementation strategy packages might also be considered. These choices are difficult because there are ever-increasing options for interventions and implementation strategies, and early decisions likely influence subsequent choices or future implementation plans (e.g., adding additional interventions in the future, when resources allow). Insurmountable barriers to implementing an intervention might arise that are unique to a given context, requiring planners to reevaluate their intervention choice, and contextually appropriate implementation strategies are challenging to select [4, 5].
In addition to considering the likely effectiveness of different intervention and/or implementation strategy combinations (hereafter referred to as decision “alternatives”), implementors also need to consider the cost implications of each alternative and determine whether they are feasible. Cost considerations—such as how much an intervention and strategy combination might cost, what the timing of those costs is, and who is responsible for those costs—have emerged as key drivers of implementation success [6,7,8,9,10,11]. During planning, implementors may ask questions such as “What might be the relative costs of implementation strategy A versus B, compared to the expected consequences of each?” or “How much staff time might be needed to implement this intervention with fidelity?” These questions about the costs and consequences of different alternatives, also called economic evaluations, can be used to help analyze trade-offs between different implementation plans [12, 13]. Economic evaluation can also help implementors plan for the financial realities of implementation and facilitate buy-in from investors or stakeholders [9, 14].
Additional decision objectives may also be important, with precise objectives being context specific. Sometimes, equity impacts are a priority, other times mitigating potential risks (e.g., harm, failing to be cost-neutral) may be important. One available resource for implementation planning that accounts for the variety of objectives and considerations of implementation is the RE-AIM project planning tool, which includes questions about the expected effects of a program, its required resources, and staff capacity . Another resource, the Implementation Research Logic Model, guides users to think through potential implementation strategies and scenarios based on known parameters . While these kinds of planning tools are extremely valuable contributions to implementation science, they are limited in that, for example, they presume a specific intervention is already chosen or provide minimal guidance on how to compare alternative implementation plans.
The complexity of implementation planning also makes existing tools limited. Thinking through implementation planning involves many characteristics that make decisions difficult: long-time horizons (requiring action up-front, though benefits may not be realized until later), the involvement of many different stakeholders with different values and preferences, uncertainty in the possible outcomes under different alternatives, and interconnected decisions [17,18,19]. This complexity makes systematic decision-making difficult, particularly because individuals tend to rely on simplifications or heuristics to make decisions in the face of complexity, leading to biased, inconsistent, or even harmful decisions [17, 20,21,22,23,24,25,26,27].
Despite the importance and complexity of implementation planning, it has received relatively little attention in the literature [3, 15, 28] and approaches are needed to help structure the planning process and weigh different alternatives. An ideal approach would be flexible enough to meet a range of planning questions and integrate multiple considerations to help answer complex questions that arise during the planning process. We believe that decision analysis, a widely used and flexible process to systematically approach decision-making, offers just that. In this paper, we discuss the decision analysis approach, with particular attention to its relevance for implementation planning questions.
What is decision analysis?
Decision analysis is a systematic way to assess various aspects of complex problems under uncertain conditions, with the goal of helping decision-makers choose the course of action (i.e., alternative) that best aligns with their objectives, considering their own context [17, 18, 29,30,31,32,33]. Applied to implementation planning questions, decision analysis aims to provide structure to the planning process by ensuring that assumptions, possible decision alternatives, available research evidence, and objectives for implementation are laid out systematically and explicitly. Within public health and healthcare, readers may be familiar with patient-level decision analysis (i.e., operationalized in clinical decision aids) that aims to help patients choose treatments that best align with their own care needs and preferences  or, with larger, national-level decision analytic approaches such as those used in the UK to structure healthcare reimbursement decisions . In the context of this paper, we take a general view and consider decision-makers to be relevant stakeholders engaged with the implementation planning processes (e.g., those who are making adoption decisions, as well as those who can choose to support/resist decisions)  and decision problems to be implementation planning questions (typically about intervention or implementation strategy choices).
We structure our discussion using the PROACTIVE framework as a guide. PROACTIVE is a foundational framework for decision analysis in health introduced by Hunink and colleagues , which draws on work from Keeney and colleagues in operations research [18, 19]. PROACTIVE offers a comprehensive overview of established steps in decision analysis while allowing for flexibility and iteration within each . The framework conceptualizes decision analysis as a process spanning three phases through which (1) the decision problem and overall objectives are defined, (2) different alternatives are compared, and (3) information on each alternative is synthesized . Overall, this framework provides a clear way to understand the full process of decision analysis, without being overly prescriptive about which specific methods are used throughout.
Throughout our discussion, we use three examples of decision analysis for implementation planning to illustrate how various steps of the PROACTIVE framework can be operationalized (Table 1). Examples were selected for their heterogenous approaches and focuses and to showcase the variety of ways that decision analysis could be approached in implementation planning efforts: selecting childhood maltreatment EBPs (Cruden et al.), improving the reach of EBPs for mental health (Zimmerman et al.), and improving rates of colorectal cancer screening (Hassmiller Lich et al.) [37,38,39].
Phase 1: PRO—defining problem and objectives
P: Defining the problem
It is critical to understand the precise nature of the problem at hand before making decisions. This often involves understanding the “natural history” of the problem—what is happening over time and what might happen if we take no action? In implementation science, this may be health problems we seek to address (e.g., disparities in blood pressure control) or low uptake of EBPs we have already decided to support (e.g., for depression treatment among veterans). Understanding the determinants of health is also important, including system structure flaws that need to be addressed as well as what barriers or facilitators to implementation exist—both broadly and specific to considered interventions. Here, it is critical that focal problems be understood in the context they will be addressed, asking and answering the question “Why does this problem persist here and now?”.
R: Reframe from other perspectives and understand objectives
Especially with complex issues in implementation that are not constrained to a single industry, discipline, or sector, multiple perspectives are critical. The problem, and what each stakeholder hopes to accomplish, might look quite different. It may take discussion to develop a shared vocabulary and understanding before the core problem, objectives, and the most critical determinants of the problem become clear. Even when prioritized decision objectives overlap, there may remain varied preferences across perspectives. Differences in objectives must be understood and acknowledged, and interconnections and commonalities highlighted to support cross-stakeholder change initiatives.
O: Focus on the unifying objective(s)
The unifying objectives of stakeholders—what a group can unite behind in a change initiative—typically become apparent as the problem is discussed and reframed. Objectives can be competing (e.g., access and system cost), and their priority could differ by stakeholders and across contexts. They might also be different over time (e.g., access matters first, then quality of care). Objectives may include meeting certain benchmarks for standard implementation outcomes such as fidelity or sustainability  or maximizing health benefits. Costs to implement and sustain the intervention are also often relevant; for example, stakeholders may want to keep costs to a minimum or within a prespecified range (potentially because of grant restrictions or a budget). Other objectives such as the potential harms of a given implementation plan or impacts on health equity may be relevant to stakeholders. Still, other objectives might be shaped by organizations’ missions. The goal in this stage of the process is to identify the set of objectives that stakeholders, as a group, agree upon in order to frame the next steps in the decision analysis process.
These first three steps are interconnected—and as such, those involved in decision analysis processes should acknowledge this and ensure these steps are flexible enough to allow for feedback and learning. The importance of stakeholder engagement in these steps cannot be overstated. Without the purposeful and meaningful engagement of relevant stakeholder groups, a problem statement and objectives might be settled on that will later be resisted—threatening sustainability—due to insufficient support or misalignment with the true structure of the system producing problematic outcomes . Traditional qualitative work (e.g., key informant interviews, focus groups) or quantitative work (e.g., surveys sent to stakeholders) can be leveraged here, and we also suggest that researchers consider methods from the field of systems science that have developed engagement processes specifically designed to facilitate a shared understanding of a problem and objectives through engaging diverse stakeholders in structured systems thinking processes—particularly valuable when motivating problems are complex [41, 42].
The case studies each approached these steps somewhat differently. First, Cruden engaged a group of stakeholders to clarify their definition of childhood maltreatment and identify objectives (referred to as criteria in their work) by which to evaluate different EBPs, such as program familiarity, strength of evidence base, or resource availability . Conversely, Zimmerman et al. began their work with a clear problem and objectives (limited reach of EBPs for mental health, increasing scheduled and completed appointments) and spent time conducting qualitative work with stakeholders to understand the components of the system under study and how perspectives on the system differed between stakeholders . Finally, Hassmiller Lich had an a priori problem definition and objectives (low rates of colorectal cancer screening, understanding the cost-effectiveness of different interventions from Medicaid’s perspective) .
Phase 2: ACT—comparing different alternatives
A: Consider relevant alternatives (interventions and/or implementation strategies)
It is critical to know the range of what alternatives are possible before decisions are made—including doing nothing (a decision itself). Information on possible interventions can come from a variety of sources, including evidence searches or consults with stakeholders and experts. Online repositories such as the National Cancer Institute’s Evidence-Based Cancer Control Programs listing can help identify interventions that align with different focus areas (e.g., HPV vaccination, tobacco control) or populations (e.g., rural adults, schoolchildren) . Tools such as the CFIR-ERIC matching tool can help narrow the possible universe of implementation strategies based on context- or intervention-specific barriers . It is also important in this stage to understand what is already in place in the community that can be built on (and not duplicated) and what resources are available to be leveraged.
C: Understand the possible consequences of each alternative
The next stage involves considering the consequences of each alternative—aligned with the objectives of interest. For example, how might the intervention impact health outcomes? What is the cost of proposed implementation strategies? Are there additional consequences that might bolster or undermine intended effects? Who might react to changes the intervention creates—and how will those reactions impact the objectives of interest?
There is a wide range of methods that can be used to understand consequences. As is true with all research, the optimal method(s) will be based on the objectives of implementation planning, available research resources and capacity, and the specific questions that need to be answered. One simple approach is to collate existing literature, reports, or evaluations on the likely consequences under each alternative. Quantitative and computational simulation modeling can be undertaken as well. Decision analysis experts including Briggs et al.  and Marshall et al. [45, 46] provide quality overviews of modeling methods and how they align with different questions of interest, along with references for additional reading. The wide range of potential methods available allows those with differing questions, expertise, resources, and/or decision urgency to engage effectively. For example, queuing and discrete event modeling are typically employed when questions focus on service delivery systems or waiting times [47, 48]. Agent-based modeling, on the other hand, can simulate interacting individuals in geographically explicit contexts, making it particularly useful when implementation planning depends on social network effects or geographic considerations .
One modeling method we wish to draw particular attention to is system dynamics modeling, which focuses on conceptualizing relationships between variables over time, identifying feedback loops between them, and modeling how the system responds to changes over time [50, 51]. These models simulate how accumulating quantities (say, the number of individuals who are up-to-date with colorectal cancer screening) change over time. As part of the broader field of systems science, system dynamics modeling has an explicit focus on and ability to simulate the elements of complexity present in implementation science work [52,53,54,55,56,57]. For example, time dynamics and delays during the implementation process are typical (e.g., change takes time, costs accrue quickly while health benefits accrue more slowly, data and system feedback is often delayed in its availability), and feedback loops can exist among relevant implementation considerations (i.e., when something in the system responds to earlier changes, either reinforcing or counteracting earlier changes—in both desired and undesired ways). When these types of characteristics are present, breaking down complex systems into pieces and simplifying assumptions allows for studying individual pathways, but evaluating pathways independently of the broader system can lead to major missed opportunities or even exacerbation of the problem that motivated intervention [23, 58, 59]. These qualities make system dynamics a natural fit for implementation science work [38, 52, 53, 60].
T: Identify and estimate trade-offs (preferences and values)
Once an understanding of the potential consequences has been established, the trade-offs between decision alternatives can be examined. For example, one implementation strategy may cost more but improve implementation outcomes more than a cheaper implementation strategy, or one implementation plan might be expected to be less effective overall but reduce health inequities. These kinds of results raise important trade-offs that decision-makers must acknowledge and consider when making decisions. To plan in the face of trade-offs requires an understanding of what decision-makers prefer and value (e.g., is it more important to improve fidelity or feasibility?), all of which may be context-specific [33, 61].
A major consideration when incorporating values and preferences into implementation planning is to consider whose values and preferences are being used [33, 62]. For example, a clinic could use the values of their patient population, or front-line providers, or administration when weighing trade-offs between alternatives. In some situations, preferences and values may align across stakeholders, and in other cases, they may not. Mixed-methods approaches can help capture how different perspectives and contexts relate to trade-offs in costs and consequences .
One particular method of use here is discrete choice experiments (DCEs), which focus on quantifying the relative importance of different aspects of alternatives . Applied to implementation planning, DCEs could be used to gather information from stakeholders on which interventions or implementation strategies they prefer , or identify trade-offs between different aspects of implementation plans (e.g., costs, effects, feasibility), allowing for a more comprehensive assessment of the value of different implementation plans.
Cruden et al. identified seven candidate EBPs for child maltreatment prevention from a best practices repository, including three in the final analyses; each intervention was scored on each stakeholder-identified criteria using published literature, and weights were used to capture preferences for different criteria . In Zimmerman’s work, stakeholders suggested many potential implementation plans, and the team worked to prioritize their top two for further analysis (implicitly incorporating preferences) . The potential effects of these plans were quantitatively compared using a system dynamics simulation model, developed and calibrated with local data (see Fig. 2 in text for a visual of the system structure) . Finally, Hassmiller Lich compared four interventions, selected via literature reviews and stakeholder interviews, using a large individual-based simulation model that estimated each intervention’s likely costs (in dollars) and effects (in years of life up-to-date on screening) .
Phase 3: IVE—synthesizing information on each alternative
I: Integrate evidence on the likely consequences given identified preferences and values
It may become clear based on anticipated consequences of alternatives and preferences/values what the “best” decision is for a specific context. Other times, it may not be. In the latter case, using a formal “objective function” that integrates objectives through weighting each component is useful. For example, quantifying the cost per additional patient reached or the net benefit of a decision alternative can integrate costs and health benefits. However, if stakeholders with different perspectives disagree on weights, it may be more useful to present information about expected outcomes for each component (i.e., valued outcome) in a tabular format (often called a “balance sheet,” “values framework,” or “performance matrix” [62, 64]).
V: Optimize the expected value
In traditional decision analysis, there are employable rules for making decisions—for example, by choosing the alternative that optimizes an objective or an objective function value or by choosing the alternative that has the smallest chance of poor outcomes. If stakeholders cannot agree on weights for each decision outcome, decision analysts might ask each to select their top 3 alternatives once outcome estimates are available or make their own weights/preferences explicit before seeing outcome estimates. In either case, a process will also be needed for determining which alternative is selected once all stakeholder selections are made (e.g., will each stakeholder get an equal vote, how will ties be broken). These decision rules should be decided through discussion with all involved stakeholders before alternatives’ scores are quantified and analyzed to minimize biases and conflicts.
E: Explore assumptions and evaluate uncertainty
Even when significant investment is made to reflect the local context in the decision analysis process when estimating potential outcomes, uncertainty will always exist. Decision analysis findings can only reflect what is known (not what is unknown), and no model will ever precisely anticipate what could be . Presenting single estimates of the consequences of alternatives can mask this uncertainty and potentially mislead decision-makers.
Once expected (best-guess) results are estimated, a cornerstone of decision analysis is exploring how consequences and trade-offs change under different assumptions. This can help decision-makers understand the likelihood of different outcomes based on the uncertainty in different aspects of the decision problem. Some uncertainties may be driven by questions about what could happen (e.g., how much an intervention would change health outcomes, or how effective an implementation strategy might be in promoting adherence). Sometimes uncertainty could be in preferences (e.g., how much is one consequence valued compared to another). Various analytic approaches can be used to explore this uncertainty and inform learning and planning. Across these approaches, we note that uncertainty can only be explored insofar as it is a known source of uncertainty, and engaging with diverse evidence bases, data, and stakeholders improves the chances that analytic models better characterize what is known and anticipate potential unknowns.
Optimization analyses help answer questions such as, “Across different implementation plans, what is the optimal course of action given specific constraints (e.g., hours available for specific types of workers, maximum allowable wait times)?” . Here, “optimal” is defined by decision-makers and often involves minimizing implementation cost or maximizing impact while not violating specific constraints . Factoring in finite resource constraints can be helpful when considering the risks and barriers related to specific implementation plans. For example, consider a clinic wanting to use grant funds to implement a new intervention, but that cannot afford the salary of a new provider. Given the varying costs/impacts for different providers (e.g., physicians, nurse practitioners) to perform implementation and intervention tasks, the clinic could identify the optimal provider (or provider mix) to undertake required tasks while avoiding hiring an additional provider.
Sensitivity analyses help identify which decision analysis model input values have the greatest leverage on outcomes [29, 65]. Single variable sensitivity analyses involve changing single inputs and recording the outcomes (i.e., deterministic sensitivity analyses). The results of these analyses can be displayed in a “tornado plot” such as Fig. 1, which clearly communicates which parameters have the potential to affect the focal outcome the most. This kind of information can refine implementation planning by helping planners understand where more (or less) attention and effort should be focused to achieve desired outcomes. Multiple inputs can also be manipulated at once to get a sense of how different scenarios impact outcomes (for example, the lowest or highest values for all inputs). Threshold conditions can also be assessed with sensitivity analyses, for example, to learn the conditions under which costs remain below a certain benchmark, or to estimate required resources to ensure a certain level of effectiveness.
Probabilistic methods can be used to quantify how uncertainty in inputs translates into holistic decision uncertainty, asking such questions as “Are conclusions about a given intervention’s benefits relative to its costs robust to simultaneous realistic variation in inputs?” . While a flavor of this type of analysis can be conducted using multivariable sensitivity analyses, this analytic approach typically involves the simulation of many combinations of inputs using probability distributions to generate a diverse, representative range of possible outcomes. This type of analysis informs decision-makers how varied outcomes might be, given plausible uncertainty in model parameters.
Uncertainty analyses can also be used to drive future research directions. Deterministic sensitivity analyses can help decision-makers pinpoint specific model uncertainties that, if addressed through further research, could improve decision-making when current uncertainty renders decision priorities unclear. Probabilistic methods to quantify uncertainty can further inform implementation research efforts through formal Value of Information analyses [12, 66,67,68]. If the overall goal is to reduce the uncertainty in a decision, then these analyses can be used to place a specific value on future research studies that can provide more information (and thus reduce decision uncertainty) [12, 66,67,68].
All case studies integrated their results. Cruden created “summary scores” for each of the three EBPs assessed . These summary scores were calculated for each stakeholder, as a weighted sum of intervention scores, using weights that stakeholders modified for their own context . Zimmerman reported visual trends in EBP reach (sessions scheduled, completed) under different implementation plans and used sensitivity analyses and stakeholder review to validate their model . Other implementation planning work, also in the VA though focused on stroke, has also used system dynamics modeling and reported extensive sensitivity and uncertainty analyses [69, 70]. Finally, Hassmiller Lich presented a visual that depicted how the cost of each intervention and the life years up-to-date were related and discussed which of the interventions were likely to be the most cost-effective, as well as the 10-year investment required . While this model was probabilistic (meaning that each replication would result in a different answer, based on initial simulated values), conducting a full probabilistic analysis was not feasible given the size of the model . Thus, they report mean values of outcomes over 10 large, full population replications .
Decision analysis provides structure to guide users to clearly articulate and agree upon objectives (e.g., improve outcomes, decrease costs, reduce inequities), uncover diverse decision alternatives, consider the strengths and weaknesses of each alternative in context, and examine potential trade-offs in objectives under each alternative . These steps are a clear fit with and can add rigor to implementation planning, where implementors typically need to compare different implementation alternatives, understand the potential consequences of those alternatives, and decide what is suitable for their context. Importantly, decision analysis does not prescribe which kinds of consequences should be examined or what is right for a given context. The choice of what consequences to assess, how to value those impacts, and how the valuation leads to a decision is always context-specific and should incorporate the preferences and values of all and often diverse stakeholders [17, 33, 61, 71]. Additionally, while many of the individual components of decision analysis may be familiar to implementation scientists (e.g., engaging stakeholders, selecting candidate implementation strategies), we believe it is valuable to situate these component pieces within a broader decision analysis framework.
All pieces of the PROACTIVE framework may not always be needed, and the component pieces within the framework may not proceed linearly. After the discussion of objectives, stakeholders might need to reevaluate the problem at hand. Sometimes, it may be that refining the problem definition and objectives provokes enough discussion to uncover a clear decision. Other times, the problem may be clear, and more effort is needed to assess the potential effects under each alternative or understand where major uncertainties are. This drives home that while the process may be constructed around coming to a decision, learning and insight gained throughout the decision analysis process can often be just as valuable [17, 71], and once a process is completed for a given decision problem, the investment made can support subsequent decision-making.
The value of applying decision analytic approaches to implementation planning has been suggested in other work, though these methods remain underutilized [12, 37, 38, 60, 71]. To advance the use of decision analysis in implementation planning and research, we propose key areas for future research and practice based on issues considered in this paper and encourage discussion on how these suggestions can be prioritized by the field.
Embrace model thinking
In addition to the complexity that can arise in the process of making decisions, complexity also exists within systems where implementation occurs (time delays in seeing the impacts of action, feedback loops, nonlinear relationships) [52,53,54,55,56]. Some modeling methods are designed to incorporate this. However, any model will, by definition, be a simplification of reality and require setting boundaries around what is included. Using initial models that represent our current, best understanding of problems and iteratively revising them as our knowledge of the system and problem grows is crucial [22, 71]. Models can be leveraged to make assumptions transparent or further improve our understanding of complex problems . For example, a small model could help identify where we need more data or knowledge—what parameters or structures are uncertain? Why are outcomes so sensitive to a specific component of the system? What happens when we consider different perspectives or constraints, or involve those with other expertise? As we learn from answers to these questions, we can expand the initial model and continue to use it to improve outcomes in complex and changing contexts.
Build the business case for decision analysis approaches
Publications and reports that describe if and how upfront investment planning improves downstream outcomes can help build momentum for future applications of decision analysis. It is possible that decision analysis could reduce long-term costs by helping implementors choose interventions and implementation strategies with the greatest chance of success in a specific context. However, we need to test this assumption in our research and invest in research that evaluates the impact of decision analysis on decision-making and downstream outcomes .
Identify when, how, and for whom decision analytic approaches are useful
As others engage in decision analysis, detailed publications of nuances, objectives, and lessons learned through the process can improve our understanding of how and when pieces of decision analysis are best employed. The application of larger modeling efforts within a decision analysis may be most helpful when many actors are involved or when large sums of money are funds at stake, like in large health systems or at the state or federal level (e.g., Hassmiller Lich et al., Table 1 ). At lower “levels” like in communities or clinics, where even more might be at stake given tighter budgets, investigating how decision analytic approaches can reasonably be deployed should be prioritized. For example, work could evaluate how a structured decision analysis process or a small model that captures the major complexities support planning efforts (e.g., Cruden et al., Table 1 ).
Improve reporting and transparency of cost data
Decision analysis approaches require estimates from published literature, drawing on, for example, studies of implementation effectiveness and work evaluating the costs of implementation strategies. Thoughtfully considering how prior literature can inform future decision-making in different contexts is thus core to any decision analysis, and all inputs used throughout a decision analysis should be interrogated and justified . One major area of current focus with implementation science that is often missing detail is the costs of implementation [74,75,76,77]. Recent publications by the “Economics and Cost” action group of the Consortium for Cancer Implementation Science have set out definitions, guidance, methods, and best practices for understanding the costs of implementation and conducting economic evaluations in implementation science [36, 74, 77,78,79,80] and complement ongoing work in the field [6, 8, 12, 14, 61, 75, 76, 81,82,83,84].
To address gaps in reporting of cost data, scholars have called for consistent and detailed reporting on intervention costs, intervention adaptations, applied implementation strategies, and accrued implementation costs [6, 9, 40, 61, 85]. Refining existing reporting guidelines to help authors standardize what is reported could greatly improve the likelihood that published work could inform decision analytic approaches in the future. For example, the Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist guides the reporting of economic evaluation [86, 87]. This checklist could be modified to prompt detailed reporting on implementation processes, including actors, timing, purpose, and costs related to each implementation decision or action—quite similar to previous calls for standardized reporting  and existing recommendations for the reporting of implementation strategies .
In addition, efforts should be made to improve the transparency of cost data and data sharing. Cost data is often sensitive or proprietary. Presenting the implementation costs associated with specific sites may allow sites to be identified, raising questions about confidentiality in research. However, incorporating costs into decision analysis depends on transparency and willingness to share data and processes, and the field should consider how to address these issues.
As an example of how we envision reporting and transparency considerations might be operationalized in future published work, Hassmiller Lich et al. specified the different cost components of each intervention they considered, whether costs were one time or recurring, and specific notes on data sources (Fig. 2) . This makes the paper of broader use even though the model was specific to North Carolina; future research can build on this work to better understand the components of costs that might be incurred if similar interventions were implemented in different contexts, even if the specific values might differ.
Increase collaborative opportunities and training
Demonstrations of decision analysis processes in the peer-reviewed literature can help bolster the evidence base for others to learn from. A focus on implementation planning and decision analysis could also be integrated into existing training programs. In situations where more expertise is needed, decision analysis experts should be engaged (perhaps specifically to help facilitate identifying a clear problem, modeling potential consequences using simulation approaches, or assessing preferences). These experts are often in disciplines like systems science, operations research, health services research, health policy, or explicit decision science fields. Many of these experts are trained to collaborate on interdisciplinary teams and can be complements to collaborators with a deep understanding of implementation complexities and subject matter expertise.
An increased attention to decision analysis can provide a dual benefit to the field of implementation science by lending structure to implementation planning and helping to uncover innovative directions for future research. A key strength of decision analysis is its flexibility to be used in the way that is best suited to a given context, and we hypothesize even a simple analysis executed thoughtfully can be powerful. We encourage implementation scientists to use decision analysis principles in their own work and report on their experiences to help drive the field forward and contribute to better implementation outcomes.
Availability of data and materials
Consolidated Framework for Implementation Research
Consolidated Health Economic Evaluation Reporting Standards
Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Adm Policy Ment Health Ment Health Serv Res. 2011;38:4–23.
Moullin JC, Dickson KS, Stadnick NA, Becan JE, Wiley T, Phillips J, et al. Exploration, Preparation, Implementation, Sustainment (EPIS) framework. Handb Implement Sci: Edward Elgar Publishing; 2020.
Moullin JC, Dickson KS, Stadnick NA, Rabin B, Aarons GA. Systematic review of the exploration, preparation, implementation, sustainment (EPIS) framework. Implement Sci. 2019;14:1.
Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Serv Res. 2017;44:177–94.
Waltz TJ, Powell BJ, Fernández ME, Abadie B, Damschroder LJ. Choosing implementation strategies to address contextual barriers: diversity in recommendations and future directions. Implement Sci. 2019;14:42.
Saldana L, Chamberlain P, Bradford WD, Campbell M, Landsverk J. The Cost of Implementing New Strategies (COINS): a method for mapping implementation resources using the stages of implementation completion. Child Youth Serv Rev. 2014;39:177–82.
Pegg SL, Walsh LM, Becker-Haimes EM, Ramirez V, Jensen-Doss A. Money makes the world go ‘round: a qualitative examination of the role funding plays in large-scale implementation and sustainment of youth evidence-based practice. Psychol Serv US: Educational Publishing Foundation. 2021;18:265–74.
Bowser DM, Henry BF, McCollister KE. Cost analysis in implementation studies of evidence-based practices for mental health and substance use disorders: a systematic review. Implement Sci. 2021;16:26.
Eisman AB, Kilbourne AM, Dopp AR, Saldana L, Eisenberg D. Economic evaluation in implementation science: making the business case for implementation strategies. Psychiatry Res. 2020;283:112433.
Ribisl KM, Leeman J, Glasser AM. Pricing health behavior interventions to promote adoption: lessons from the marketing and business literature. Am J Prev Med. 2014;46:653–9 2014/05/21 ed.
Roberts SLE, Healey A, Sevdalis N. Use of health economic evaluation in the implementation and improvement science fields—a systematic literature review. Implement Sci. 2019;14:72.
Hoomans T, Severens JL. Economic evaluation of implementation strategies in health care. Implement Sci. 2014;9:168.
Raghavan R. The role of economic evaluation in dissemination and implementation research. Dissem Implement Res Health. 2nd ed. New York: Oxford University Press; 2017.
Dopp AR, Narcisse M-R, Mundey P, Silovsky JF, Smith AB, Mandell D, et al. A scoping review of strategies for financing the implementation of evidence-based practices in behavioral health systems: state of the literature and future directions. Implement Res Pract. 2020;1:2633489520939980.
Use When Planning a Project – RE-AIM. [cited 2022 Apr 14]. Available from: https://re-aim.org/applying-the-re-aim-framework/re-aim-guidance/use-when-planning-a-project/
Smith JD, Li DH, Rafferty MR. The implementation research logic model: a method for planning, executing, reporting, and synthesizing implementation projects. Implement Sci. 2020;15:84.
Hunink MGM, Weinstein MC, Wittenberg E, Drummond MF, Pliskin JS, Wong JB, et al. Decision making in health and medicine: integrating evidence and values. 2nd ed. Cambridge: Cambridge University Press; 2014.
Keeney RL. Decision analysis: an overview. Oper Res INFORMS. 1982;30:803–38.
Hammond JS, Keeney RL, Raiffa H. Smart choices: a practical guide to making better decisions. Boston: Harvard Business Review Press; 2015.
Kahneman D, Tversky A. Prospect theory: an analysis of decision under risk. Econometrica. 1979;47:263–91 [Wiley, Econometric Society].
Redelmeier DA, Shafir E. Medical decision making in situations that offer multiple alternatives. JAMA. 1995;273:302–5.
Lemke MK. Model thinking and formal modeling to improve our mental models in population health research. Complex Syst Popul Health. New York: Oxford University Press; 2020.
Sterman JD. Learning from evidence in a complex world. Am J Public Health. 2006;96:505–14.
Sterman JD. Modeling managerial behavior: misperceptions of feedback in a dynamic decision making experiment. Manag Sci INFORMS. 1989;35:321–39.
Tversky A, Kahneman D. The framing of decisions and the psychology of choice. Science. 1981;211:453–8 American Association for the Advancement of Science.
Redelmeier DA, Rozin P, Kahneman D. Understanding patients’ decisions. Cognitive and emotional perspectives JAMA. 1993;270:72–6.
Tversky A, Kahneman D. Judgment under uncertainty: heuristics and biases. Science. 1974;185:1124–31 American Association for the Advancement of Science.
Moullin JC, Sabater-Hernández D, Fernandez-Llimos F, Benrimoj SI. A systematic review of implementation frameworks of innovations in healthcare and resulting generic implementation framework. Health Res Policy Syst. 2015;13:16.
Briggs A, Sculpher M, Claxton K. Decision modelling for health economic evaluation. Oxford: OUP Oxford; 2006.
Ragsdale C. Spreadsheet modeling and decision analysis: a practical introduction to business analytics. Boston: Cengage Learning; 2014.
Parnell GS, Terry Bresnick MBA, Steven N. Tani PhD, Eric R. Johnson PhD. Handbook of decision analysis. Hoboken: Wiley; 2013.
Ozcan YA. Quantitative methods in health care management: techniques and applications. San Francisco: John Wiley & Sons; 2005.
Sheldrick RC, Hyde J, Leslie LK, Mackie T. The debate over rational decision making in evidence-based medicine: implications for evidence-informed policy. Evid Policy. 2021;17:147–59 Bristol, United Kingdom: Policy Press.
Witteman HO, Ndjaboue R, Vaisson G, Dansokho SC, Arnold B, Bridges JFP, et al. Clarifying values: an updated and expanded systematic review and meta-analysis. Med Decis Mak Int J Soc Med Decis Mak. 2021;41:801–20.
Baltussen R, Marsh K, Thokala P, Diaby V, Castro H, Cleemput I, et al. Multicriteria decision analysis to support health technology assessment agencies: benefits, limitations, and the way forward. Value Health J Int Soc Pharmacoeconomics Outcomes Res. 2019;22:1283–8.
Eisman AB, Quanbeck A, Bounthavong M, Panattoni L, Glasgow RE. Implementation science issues in understanding, collecting, and using cost estimates: a multi-stakeholder perspective. Implement Sci. 2021;16:75.
Cruden G, Frerichs L, Powell BJ, Lanier P, Brown CH, Lich KH. Developing a multi-criteria decision analysis tool to support the adoption of evidence-based child maltreatment prevention programs. Prev Sci Off J Soc Prev Res. 2020;21:1059–64.
Zimmerman L, Lounsbury DW, Rosen CS, Kimerling R, Trafton JA, Lindley SE. Participatory system dynamics modeling: increasing stakeholder engagement and precision to improve implementation planning in systems. Adm Policy Ment Health Ment Health Serv Res. 2016;43:834–49.
Hassmiller Lich K, Cornejo DA, Mayorga ME, Pignone M, Tangka FK, Richardson LC, et al. Cost-effectiveness analysis of four simulated colorectal cancer screening interventions, North Carolina. Prev Chronic Dis. 2017;14:E18.
Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. 2013;8:139.
Andersen DF, Vennix JA, Richardson GP, Rouwette EA. Group model building: problem structuring, policy simulation and decision support. J Oper Res Soc. 2007;58:691–4.
Hovmand PS. Group model building and community-based system dynamics process. Community Based Syst Dyn. New York: Springer; 2014. p. 17–30.
Evidence-based programs listing | evidence-based cancer control programs (EBCCP). [cited 2022 Mar 30]. Available from: https://ebccp.cancercontrol.cancer.gov/searchResults.do
Briggs ADM, Wolstenholme J, Blakely T, Scarborough P. Choosing an epidemiological model structure for the economic evaluation of non-communicable disease public health interventions. Popul Health Metr. 2016;14:17.
Marshall DA, Burgos-Liz L, Ijzerman MJ, Osgood ND, Padula WV, Higashi MK, et al. Applying dynamic simulation modeling methods in health care delivery research—the SIMULATE Checklist: report of the ISPOR Simulation Modeling Emerging Good Practices Task Force. Value Health. 2015;18:5–16.
Marshall DA, Burgos-Liz L, IJzerman MJ, Crown W, Padula WV, Wong PK, et al. Selecting a dynamic simulation modeling method for health care delivery research-part 2: report of the ISPOR Dynamic Simulation Modeling Emerging Good Practices Task Force. Value Health J Int Soc Pharmacoeconomics Outcomes Res. 2015;18:147–60.
Fuller D, Sinclair E, Lamb HR, Cayce JD, Snook J. Emptying the new asylums: a beds capacity model to reduce mental illness behind bars. Arlington, VA: Treatment Advocacy Center; 2017. Available from: https://www.treatmentadvocacycenter.org/storage/documents/emptying-new-asylums.pdf.
La EM, Lich KH, Wells R, Ellis AR, Swartz MS, Zhu R, et al. Increasing access to state psychiatric hospital beds: exploring supply-side solutions. Psychiatr Serv Wash DC. 2016;67:523–8.
Nowak SA, Parker A, Radhakrishnan A, Schoenborn N, Pollack CE. Using an agent-based model to examine de-implementation of breast cancer screening. Med Care. 2021;59:e1-8.
Homer JB, Hirsch GB. System dynamics modeling for public health: background and opportunities. Am J Public Health. 2006;96:452–8.
Lich KH, Ginexi EM, Osgood ND, Mabry PL. A call to address complexity in prevention science research. Prev Sci. 2013;14:279–89.
Burke JG, Lich KH, Neal JW, Meissner HI, Yonas M, Mabry PL. Enhancing dissemination and implementation research using systems science methods. Int J Behav Med. 2015;22:283–91.
Luke DA, Morshed AB, McKay VR, Combs TB. Systems science methods in dissemination and implementation research. Dissem Implement Res Health. 2nd ed. New York: Oxford University Press; 2017.
Hassmiller Lich K, Frerichs L, Fishbein D, Bobashev G, Pentz MA. Translating research into prevention of high-risk behaviors in the presence of complex systems: definitions and systems frameworks. Transl Behav Med. 2016;6:17–31.
Northridge ME, Metcalf SS. Enhancing implementation science by applying best principles of systems science. Health Res Policy Syst. 2016;14:74.
Holmes B, Finegood D, Riley B, Best A. Systems thinking in dissemination and implementation research. In Dissemination and Implementation Research in Health: Translating Science to Practice. Oxford: Oxford University Press; 2012. Retrieved from https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199751877.001.0001/acprof-9780199751877-chapter-9. Accessed 19 July 2022.
Ornstein JT, Hammond RA, Padek M, Mazzucca S, Brownson RC. Rugged landscapes: complexity and implementation science. Implement Sci. 2020;15:85.
Apostolopoulos Y. Bridging the divide. Complex Syst Popul Health. 2020.
Patterson MS, Lemke MK, Nelon J. Complex systems in a nutshell. Complex Syst Popul Health. New York: Oxford University Press; 2020. p. 3-19.
Atkinson J-A, Page A, Wells R, Milat A, Wilson A. A modelling tool for policy analysis to support the design of efficient and effective policy responses for complex public health problems. Implement Sci. 2015;10:26.
Dopp AR, Mundey P, Beasley LO, Silovsky JF, Eisenberg D. Mixed-method approaches to strengthen economic evaluations in implementation research. Implement Sci. 2019;14:2.
Thokala P, Devlin N, Marsh K, Baltussen R, Boysen M, Kalo Z, et al. Multiple criteria decision analysis for health care decision making–an introduction: report 1 of the ISPOR MCDA Emerging Good Practices Task Force. Value Health J Int Soc Pharmacoeconomics Outcomes Res. 2016;19:1–13.
Salloum RG, Shenkman EA, Louviere JJ, Chambers DA. Application of discrete choice experiments to enhance stakeholder engagement as a strategy for advancing implementation: a systematic review. Implement Sci. 2017;12:140.
Dolan JG, Veazie PJ. Balance sheets versus decision dashboards to support patient treatment choices: a comparative analysis. The patient. 2015;8:499–505.
Drummond MF, Sculpher MJ, Claxton K, Stoddart GL, Torrance GW. Methods for the economic evaluation of health care programmes. Oxford: Oxford University Press; 2015.
Fenwick E, Claxton K, Sculpher M. The value of implementation and the value of information: combined and uneven development. Med Decis Mak. 2008;28:21–32 2008/02/12 ed.
Heath A, Kunst N, Jackson C, Strong M, Alarid-Escudero F, Goldhaber-Fiebert JD, et al. Calculating the expected value of sample information in practice: considerations from 3 case studies. Med Decis Making. 2020;40:314–26.
Kunst N, Wilson ECF, Glynn D, Alarid-Escudero F, Baio G, Brennan A, et al. Computing the expected value of sample information efficiently: practical guidance and recommendations for four model-based methods. Value Health. 2020;23:734–42.
Tian Y, Hassmiller Lich K, Osgood ND, Eom K, Matchar DB. Linked sensitivity analysis, calibration, and uncertainty analysis using a system dynamics model for stroke comparative effectiveness research. Med Decis Making. 2016;36:1043–57 SAGE Publications Inc STM.
Lich KH, Tian Y, Beadles CA, Williams LS, Bravata DM, Cheng EM, et al. Strategic planning to reduce the burden of stroke among veterans: using simulation modeling to inform decision making. Stroke. 2014;45:2078–84.
Sheldrick RC, Cruden G, Schaefer AJ, Mackie TI. Rapid-cycle systems modeling to support evidence-informed decision-making during system-wide implementation. Implement Sci Commun. 2021;2:116.
Cilenti D, Issel M, Wells R, Link S, Lich KH. System dynamics approaches and collective action for community health: an integrative review. Am J Community Psychol. 2019;63:527-45.
Eddy DM, Hollingworth W, Caro JJ, Tsevat J, McDonald KM, Wong JB. Model transparency and validation: a report of the ISPOR-SMDM Modeling Good Research Practices Task Force-7. Value Health. 2012;15:843–50.
Gold HT, McDermott C, Hoomans T, Wagner TH. Cost data in implementation science: categories and approaches to costing. Implement Sci. 2022;17:11.
Cidav Z, Mandell D, Pyne J, Beidas R, Curran G, Marcus S. A pragmatic method for costing implementation strategies using time-driven activity-based costing. Implement Sci. 2020;15:28.
Huebschmann AG, Trinkley KE, Gritz M, Glasgow RE. Pragmatic considerations and approaches for measuring staff time as an implementation cost in health systems and clinics: key issues and applied examples. Implement Sci Commun. 2022;3:44.
O’Leary MC, Hassmiller Lich K, Frerichs L, Leeman J, Reuland DS, Wheeler SB. Extending analytic methods for economic evaluation in implementation science. Implement Sci. 2022;17:27.
Dopp AR, Kerns SEU, Panattoni L, Ringel JS, Eisenberg D, Powell BJ, et al. Translating economic evaluations into financing strategies for implementing evidence-based practices. Implement Sci. 2021;16:66.
Barnett ML, Stadnick NA, Proctor EK, Dopp AR, Saldana L. Moving beyond Aim Three: a need for a transdisciplinary approach to build capacity for economic evaluations in implementation science. Implement Sci Commun. 2021;2:133.
Saldana L, Ritzwoller DP, Campbell M, Block EP. Using economic evaluations in implementation science to increase transparency in costs and outcomes for organizational decision-makers. Implement Sci Commun. 2022;3:40.
Michaud TL, Hill JL, Heelan KA, Bartee RT, Abbey BM, Malmkar A, et al. Understanding implementation costs of a pediatric weight management intervention: an economic evaluation protocol. Implement Sci Commun. 2022;3:37.
Wagner TH. Rethinking how we measure costs in implementation research. J Gen Intern Med. 2020;35:870–4.
Wagner TH, Dopp AR, Gold HT. Estimating downstream budget impacts in implementation research. Med Decis Mak. 2020;40:968–77 2020/09/22 ed.
Wagner TH, Yoon J, Jacobs JC, et al. Estimating Costs of an Implementation Intervention. Medical Decision Making. 2020;40(8):959-67.
Wiltsey Stirman S, Baumann AA, Miller CJ. The FRAME: an expanded framework for reporting adaptations and modifications to evidence-based interventions. Implement Sci. 2019;14:58.
Husereau D, Drummond M, Petrou S, Carswell C, Moher D, Greenberg D, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. J Med Econ. 2013;16:713–9.
Husereau D, Drummond M, Augustovski F, de Bekker-Grob E, Briggs AH, Carswell C, et al. Consolidated Health Economic Evaluation Reporting Standards 2022 (CHEERS 2022) Statement: updated reporting guidance for health economic evaluations. Appl Health Econ Health Policy. 2022;20:213–21.
We would like to acknowledge the Economics and Implementation Science Workgroup for their formative input and internal review process. We would especially thank Meghan O’Leary, Dr. Alex Dopp, and Dr. Enola Proctor for their feedback on the early drafts of the manuscript.
This project was supported by NIH grant numbers T32HD091058 and T32CA057711. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIH.
Ethics approval and consent to participate
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Smith, N.R., Knocke, K.E. & Hassmiller Lich, K. Using decision analysis to support implementation planning in research and practice. Implement Sci Commun 3, 83 (2022). https://doi.org/10.1186/s43058-022-00330-1
- Implementation science
- Cost and cost analysis
- Economic evaluation
- Decision support techniques