Skip to main content

Understanding implementation costs of a pediatric weight management intervention: an economic evaluation protocol

Abstract

Background

Understanding the cost and/or cost-effectiveness of implementation strategies is crucial for organizations to make informed decisions about the resources needed to implement and sustain evidence-based interventions (EBIs). This economic evaluation protocol describes the methods and processes that will be used to assess costs and cost-effectiveness across implementation strategies used to improve the reach, adoption, implementation, and organizational maintenance of an evidence-based pediatric weight management intervention- Building Health Families (BHF).

Methods

A within-trial cost and cost-effectiveness analysis (CEA) will be completed as part of a hybrid type III effectiveness-implementation trial (HEI) designed to examine the impact of an action Learning Collaborative (LC) strategy consisting of network weaving, consultee-centered training, goal-setting and feedback, and sustainability action planning to improve the adoption, implementation, organizational maintenance, and program reach of BHF in micropolitan and surrounding rural communities in the USA, over a 12-month period. We discuss key features of implementation strategy components and the associated cost collection and outcome measures and present brief examples on what will be included in the CEA for each discrete implementation strategy and how the results will be interpreted. The cost data will be collected by identifying implementation activities associated with each strategy and using a digital-based time tracking tool to capture the time associated with each activity. Costs will be assessed relative to the BHF program implementation and the multicomponent implementation strategy, included within and external to a LC designed to improve reach, effectiveness, adoption, implementation, and maintenance (RE-AIM) of BHF. The CEA results will be reported by RE-AIM outcomes, using the average cost-effectiveness ratio or incremental cost-effectiveness ratio. All the CEAs will be performed from the community perspective.

Discussion

The proposed costing approach and economic evaluation framework for dissemination and implementation strategies and EBI implementation will contribute to the evolving but still scant literature on economic evaluation of implementation and strategies used and facilitate the comparative economic analysis.

Trial registration

ClinicalTrials.gov NCT04719442. Registered on January 22, 2021.

Background

Childhood obesity is associated with increased risk for a variety of cardiovascular and other diseases and has been documented to predict adult obesity which results in personal health challenges and imposes an economic burden to the healthcare system [1, 2]. Although pediatric weight management interventions (PWMIs) or prevention programs are becoming prevalent, families from micropolitan (i.e., cities <50,000 populations) and surrounding rural areas have limited access to PWMIs that are typically offered in larger urban areas. Expanding the availability of evidence-based PWMIs and increasing capacity in micropolitan regions to implement these interventions can have a significant public health impact. Systems-based approaches that engage multiple community-organizational partners to share implementation responsibility have been proposed as a potential pathway to increasing PWMI access in micropolitan communities [3].

Integrated system-based approaches typically include horizontal components (i.e., engagement of organizations across a community) and vertical components (i.e., engagement of administrative decision makers and staff that would ultimately implement the PWMI) [4]. Both horizontal and vertical system components provide opportunities for applying dissemination and implementation (D&I) strategies that may improve the likelihood that evidence-based PWMIs can be adapted, adopted, implemented, and sustained in micropolitan areas [3, 5]. For example, a fund and contract strategy, involving research funding announcement, application review, and contractual agreements for implementation, can be used to improve the adoption of PWMIs in communities with the readiness for implementation [6]. Moreover, the use of an action learning collaborative (LC) strategy that encompasses a variety of activities to support communities in improving PWMI reach, effectiveness, adoption, implementation, and maintenance (RE-AIM [7]) (e.g., structured learning/training sessions, goal-setting and feedback, sustainability action planning) can be used to increase community engagement across organizations and facilitate capacity building for PWMI implementation [8, 9].

Despite the proposed promise of these approaches, it is necessary to understand both the ability of these strategies to improve RE-AIM outcomes as well as the costs and resources needed to execute D&I strategies [10,11,12]. Unfortunately, study findings from those that have conducted economic evaluations of D&I strategies are mixed and thus the economic feasibility of D&I strategy use remain unclear. In studies that have applied economic analysis of LCs [13, 14], a cost-effectiveness analysis (CEA) focusing on pre-post changes in clinician’s competency on trauma-focused cognitive behavioral therapy and youth’s trauma-related mental health found the LC did not produce a favorable cost-effectiveness ratio with clinician competency as the outcomes measure. However, the subsequent reductions in youth psychopathology demonstrated a high cost-effectiveness. More broadly, two recent review studies [15, 16] identified articles examining the costs, consequences, and cost-effectiveness of strategies designed to influence the adoption of public or population-level interventions in community settings. Both reviews shared the methodology limitation that the number of studies that they identified and synthesized, especially among studies associated with the implementation or implementation strategies, and a plethora of terms and definitions of D&I strategies making cross-study comparisons related to CEA challenging. Moreover, no detailed discussion of costing methods was found in their reviews.

As a result of the challenge of summarizing CEAs relative to D&I strategies, there is a movement to increase the specification of both D&I strategies and the resulting economic evaluation [17]. The recent effort made by the “Economics and Cost” action group of the Consortium for Cancer Implementation Science is an example of this movement. In a series of papers [10,11,12, 18], this effort provided common definitions, costing methods, developing costing guidance, and guidance to build collaborations across disciplines [19]. Complementary to the collection of articles, this paper provides a description of an economic evaluation across various D&I outcomes using RE-AIM [7] and methods to cost D&I strategies mapped to specific outcomes related to PWMI implementation in micropolitan settings. Specifically, it outlines the methods and processes of the prospective within-trial economic evaluation of D&I strategies to be undertaken within a hybrid type III effectiveness-implementation (HEI) trial to improve RE-AIM outcomes of evidence-based PWMIs in micropolitan communities. We will examine costs and cost-effectiveness of D&I strategies focused on enhancing PWMI (1) reach (i.e., the number, proportion, and representativeness of program participants relative to the intended audience), (2) adoption (i.e., the number and proportion of communities that agree to deliver the PWMI), (3) implementation (i.e., adherence to program protocol), (4) organizational maintenance (i.e., sustained implementation to addition cohorts of families over time), and (5) their overall impact on the effectiveness (i.e., reduction in BMI z scores). The cost-effectiveness results will provide data for payers, policymakers, and providers to make informed decisions about specific strategies for specific RE-AIM outcome targets, especially in organizations or communities where resources are scarce.

Methods

Study design

This economic evaluation protocol is embedded within a type III HEI trial, aiming to test multilevel D&I strategies that focus on the RE-AIM outcomes of an evidence-based PWMI (i.e., Building Health Families [BHF] [20,21,22]). Details of the overarching trial have been published elsewhere [3, 5, 23] and information on that trial is presented here as it relates to the economic evaluation of strategies that target RE-AIM outcomes (see Table 1). Specifically, the trial includes three primary D&I strategies—fund and contract, an online, packaged implementation blueprint, and an action LC. In addition, the implementation blueprint includes activities focused on training, program implementation, and referral resources to improve program reach. Similarly, the LC includes activities focused on network weaving, used to promote information sharing within and outside the organizations building upon existing working relationships [6], consultee-centered training, and sustainability action planning. The study is intended to (1) identify local demand for BHF, a PWMI that was used as the basis for the online, packaged implementation blueprint (referred to hereafter as the BHF Online Training Resources and Program Package [5]); (2) examine potential determinants of BHF adoption; and (3) compare BHF RE-AIM outcomes in communities that participate in the LC to those that do not. As a pilot type III HEI, the primary aim is to determine if the LC communities document better D&I outcomes (reach, adoption, implementation, BHF sustainability) and the secondary aim is to determine the relative effectiveness of BHF between communities that participate in the LC versus those that do not. Of note, regardless of LC participation status, communities are exposed to strategies of fund and contract and BHF Online Training Resources and Program Package. The trial protocol is registered with clinicaltrials.gov (NCT04719442).

Table 1 Cost and outcome measures for the within-trial economic evaluation by the RE-AIM outcomes

Each of the D&I strategies is described below. To ensure clear communication and transparency, we used Proctor and colleagues’ framework [24] to specify implementation strategies. The specification of each D&I strategy by the actor (who enacts the strategy?), the action(s) (what are the specific actions, steps, or processes that need to be enacted?), the action target (what constructs are targeted? What is the unit of analysis?), temporality (when is the strategy used?), dose (what is the intensity?), and implementation outcome (what implementation outcome(s) are likely to be affected by each strategy?) is presented in Table 2. We also individually assess the costs of each strategy.

Table 2 Labor activity specification of dissemination and implementation strategies used in the HEI trial by the Proctor framework [24]

Fund and contract dissemination strategy focused on adoption

Fund and contract dissemination strategies typically include a request for applications (RFA) to deliver an evidence-based intervention, contracting to ensure understanding of what delivery will entail, and some sort of funding to improve local motivation to adopt the given intervention [6]. As part of this trial, the fund and contract dissemination strategy is used to identify communities with strong cross-organization partnerships and a priority to address childhood obesity locally. Table 2 includes a full specification of this strategy and the HEI trial includes 2 phases—a call for letters of intent (LOI) and an invitation for full proposals based on LOI review. Both processes are developed by the research team and reviewed and revised in partnership with the community advisory board (BHF-CAB). The call for LOIs focuses on describing the attributes of the BHF Online Training Resources and Program Package (i.e., relative advantage, compatibility, complexity, observability, and trialability [25]). The call also highlights financial resources available for the community and the need for a horizontal (i.e., across community organizations) and vertical (i.e., decision makers and doers) systems-based approach to BHF implementation. The target of the fund and contract strategy was to increase community adoption of BHF [23].

Developing packaged implementation blueprint: BHF Online Training Resources and Program Package focused on adoption and implementation fidelity

A detailed description of the BHF Online Training Resources and Program Package has been published previously [5]. In brief, the package was designed for BHF community facilitators to provide all of the training and materials necessary for program implementation. Specific training modules were developed for a program facilitator and for facilitators of the specific content areas (nutrition, physical activity, and lifestyle). Introductory modules also focused on program initiation and methods to improve program reach. Modules included presentation materials, handouts, lesson plans, facilitator knowledge checks, and self-assessed implementation fidelity checklists. To allow the program facilitator to track family progress and implementation fidelity a data portal was included in the package. The primary target of the BHF Online Training Resources and Program Package was to increase implementation fidelity (operationalized as adherence to protocol and increased facilitator capacity) while allowing adaptability of delivery within and across core intervention elements [26]. This strategy has secondary targets of improved adoption through reducing implementation complexity and improved reach.

Building Health Families Action Learning Collaborative (BHF-LC) to improve implementation quality

The BHF-LC represents the primary independent variable for the HEI trial whereby 2 to 4 communities receive only the BHF Online Training Resources and Program Package (package only, BHF-PO) and 2 to 4 communities will receive the package and participate in a LC (BHF-LC). The intent of a LC is to bring together community teams with members from different organizations and professional roles (e.g., staff, supervisors, senior leaders) to work together to learn BHF and sustain its use over time. Activities used in the LC model include pre-work activities to build capacity, in-person trainings, and implementation periods between trainings that involve ongoing consultation and quality improvement strategies [6]. Specifically, the BHF-LC includes collaborative learning sessions followed by action periods over a 2-year implementation period. The learning session topics align with the timing of BHF program activities with an early focus on reach and recruitment, followed by a focus on implementation quality, followed by a focus on maintenance of community program implementation. Consultee-centered training [27] is embedded within learning sessions to improve facilitator implementation capacity. This includes instruction on core content, review of BHF implementation data, facilitator reflection on implementation processes, goal setting and feedback relative to implementation outcomes, and group-based problem-solving to improve reach, implementation fidelity, and sustainability. Finally, sustainability action planning will be included in the BHF-LC and focus on strategies to embed program implementation within organization workflows and job responsibilities. The sustainability action planning will also include engaging local payers and public health organizations to explore opportunities for ongoing funding and reimbursement for implementation.

Evidence-based pediatric weight management intervention

BHF is a 12-month parent-child dyad PWMI program, adapted from Epstein and colleagues’ efficacious traffic light diet (TLD) PWMI, and was tailored to fit rural/micropolitan communities [20, 21]. The BHF program includes a minimum of 32 contact hours consisting of three main program components: nutrition education, behavior modification, and physical activity. Education is provided to children and parents together and independently based on the topic and depth of information [20, 21]. Participants and parents are expected to attend 12 continuous weeks of education (2 h/session) followed by 12 weeks of relapse prevention refresher courses (1 h per session for every 3 weeks). One additional refresher session (2 h) is conducted between 6 months and 1 year to revisit the TLD eating plan and environmental changes, re-ignite social support within the group, and finish with a fun activity. Two follow-up check-in sessions are conducted at 6 and 12 months for approximately 1 h per session.

Eligibility, recruitment, and allocation

Community level

Applying the community engagement approach, the research team and the BHF-CAB disseminated the RFA within and outside their organizations and networks (i.e., applying network weaving strategy) and documented that communities across all 93 counties received the RFA. Eight communities with broad geographic distribution responded to the RFA of which seven rural communities in Nebraska completed a full proposal and were selected as the participating communities for the HEI trial. We then ranked the seven applications in a descending order and allocated the odd-numbered rankings to the (BHF-LC) and the even-numbered rankings to the package-only (BHF-PO) condition [23]. This exercise resulted in four communities in the BHF-LC condition and three communities in the BHF-PO condition. All communities that implement the PWMI will have access to the BHF Online Training Resources and Program Package that includes all of the materials necessary to implement and evaluate the BHF intervention.

Communities assigned to the BHF-PO condition are required to complete a memorandum of agreement (i.e., obtaining formal commitment strategy) to report on the implementation of the PWMI and to use pragmatic evaluation strategies for reach and effectiveness outcomes and to provide de-identified program data for baseline, 6, and 12 months. These communities will also have technical support as needed related to package function, but no other facilitation. Communities allocated to the BHF-LC condition will participate in LC training lessons with additional embedded implementation support activities with a goal to (1) develop local strategies to screen and engage families through the BMI Reporter in schools or clinical screening and referral, (2) increase implementation fidelity, and (3) plan for sustainability.

Individual level

Recruitment will be completed by selected communities through school BMI report cards, physician referral, local advertisements, and through word-of-mouth based on locally available resources and guided by the BHF Online Training Resources and Program Package. All materials are designed for local adaptation based on location, time of the session, clinical and school partners, and form of recruitment used. For BHF participation, families will be eligible if they have a 6- to 12-year-old child with a BMI percentile ranking at or above the 95th percentile. Children with major cognitive or physical impairments, parents or children with a contraindication for physical activity and families participating in a concurrent PWMI will be excluded. According to our previous studies [20,21,22], 8–12 families per cohort would be ideal with a minimum of 5 families and a maximum of 12 families in each cohort. The study goal is to enroll 12 families (n=30–48 individuals) per cohort and 2 cohorts per community. With a total of 7 communities recruited, the total PWMI intervention participants are approximately 420–672 parent-child dyads.

Overview of economic evaluation

This protocol has been reported in line with the Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist [28] (see Supplementary File 1: CHEERS Checklist).

The economic evaluation of the HEI trial is two-fold. First is to conduct a cost and cost-effectiveness analysis of the fund and contract dissemination strategy for the BHF adoption, the BHF Online Training Resources and Program Package relative to BHF adoption and implementation, and the BHF-LC implementation strategy for promoting BHF adoption, implementation fidelity, and sustainability for the duration of the trial (12 months). The second is to conduct a cost and cost-effectiveness analysis of the BHF intervention. The CEA will be conducted from an implementing community [29] and limited societal perspective (i.e., including only productivity loss of participants due to the participation in the BHF program). The perspective of implementing community would include costs of time spent on planning, training, recruitment, delivering, and post-session follow-up, resources used, communication with the research team and BHF participants, and opportunity costs of community facilitators. Moreover, we will include the time and costs of the research team for providing training, facilitation, and technical assistance. Costs associated with research administration, data collection, and trial outcome assessment will be excluded. All costs will be categorized as labor and non-labor costs (e.g., space, supplies, or information technology) and expressed as US $2022. The unit of analysis for the economic evaluation is the 7 communities consisting of approximately 21–28 community facilitators who will be provided with BHF Online Training Resources and Program Package with or without a LC implementation strategy to implement and deliver the BHF intervention. The time horizon of the within-trial CEA will be across a 15-month period of time (first 3 months for preparation, and later 12 months for BHF intervention), to incorporate sufficient time to complete the whole implementation process of one BHF cohort, initiating from preparation, training, recruitment, and program delivery to post-intervention follow-up. This differs from the primary clinical effectiveness time horizon of 6 months. Costs and benefits will not be discounted due to a short duration.

Effect measurement

The outcomes measured across implementation strategies and BHF intervention are described in Table 1. At the community level, the outcome associated with the fund and contract dissemination strategy is BHF program adoption, defined as (1) an objective count of communities responding to call for LOI and full application and (2) an objective count of communities responding to RFA and contracted to implement BHF. In addition, the BHF Online Training Resources and Program Package strategy targets outcomes of adoption, measured as an objective count of numbers of BHF cohorts initiated, and implementation, operationalized as community facilitator competence using the knowledge check score obtained from the BHF online training resource and direct observation of implementation fidelity (i.e., adherence to protocol, engagement, and dose [i.e., the number and duration of sessions conducted per cohort]). Finally, the LC implementation strategy involves outcomes of (1) reach, measured as the number, proportion, and representativeness of eligible family screened; (2) adoption, measured as an objective count of BHF cohorts initiated; (3) implementation fidelity, measured using the fidelity checklist via direct observation conducted by the trained research staff; and (4) sustainability, operationalized as an objective count of BHF cohort continued following the completion of LC sessions.

At the individual level, the effectiveness outcome will be assessed using a matched cohort design to determine changes in BMI z-scores from baseline to 6 and 12 months among BHF intervention participants across BHF-LC and BHF-PO conditions. We will apply an intention-to-treat approach when conducting the economic evaluation of the BHF intervention (effectiveness as the outcome measure).

Cost measurement

Labor costs

Dissemination and implementation strategies

The cost estimation of D&I strategies is guidied by the activity-based costing and Proctor framework, which is similar to the modified time-driven activity-based costing (TDABC) proposed by Cidav and colleagues [30]. We further extend the activity-based costing approach of implementation strategies to include opportunity costs, which are often overlooked and are key elements when considering the perspective of small-scale stakeholders [29], incurred by individuals (in our case, community facilitators) who are the target of D&I strategies. As the costs of D&I strategies are typically labeled as either labor (from performing the implementation activities) or non-labor costs (derived from the invoice or administrative data), the labor costs in a D&I strategy will be calculated by multiplying the hourly wages of personnel (actor) who conduct activities (action) associated with a D&I strategy with the amount of times and frequencies (dose and temporality) spent on activities plus the opportunity costs incurred by the strategy recipients (action target) (Table 2). The costs per D&I strategy will be calcuated as: \({C}_i={S}_i+\sum_{j=1}^m{X}_{ij}\), where Ci = costs per discrete D&I strategy i; Xij=costs associated with activity j assigned to a D&I strategy i; j = activity j associated with a D&I strategy i; j = 1, 2, 3, …, m; and i =1,2,3,...; and Si = amount of non-labor costs assigned to implementation strategy i. We will use the national hourly wage rate, corresponding to the position identified, of each community facilitator, CAB members, and research team members for estimates of labor costs.

Table 2 details the specification of each D&I strategy used in the HEI trial and the application of the specification for estimates of labor costs. For example, categorized by implementation activities, the actor in the fund and contract dissemination strategy includes research team members, CAB members, and a project coordinator. The hourly wage rate of each personnel involved in a specific activity can be easily derived from the the US Bureau of Labor Statistics [31]. The activities involved in this strategy including drafting the LOI and the full applications to be distributed to interested communities that may respond to the RFA, and reviewing the application materials, meetings to discuss the application and make a decision, and contacting the awardees. To obtain the total cost for each implementation activity, we will multiply the total time spent on each activity by each research team and CAB member and community facilitator by their corresponding hourly wage rate. Similar exercises will be applied to all other activities. The total labor cost of a discrete D&I strategy will be derived by summing up the total cost of each implementation activity associated with that discrete strategy. Similarly, we will calculate the total labor cost of the LC strategy and the BHF Online Training Resources and Program Package by summing up the costs associated with each activity specified (Table 2), applying the similar steps stated above.

BHF program training, recruitment, and delivery

The cost of the BHF program is estimated by the following components—BHF online module training, participant recruitment, and BHF program delivery. We will use the similar costing approach proposed for D&I strategies to estimate the costs of BHF program training, recruitment, and delivery. The labor cost of the BHF program training, recruitment, and delivery will be estimated by the time spent on activities of training, participant recruitment, session preparation, delivery, and post-session follow-up and national hourly wage rates of the similar position of community facilitators who enacted the activities. In the limited societal perspective, we will take into account the opportunity cost of the families attended the intervention sessions and other use (e.g., individual consultation), using the average national average hourly wage rate [31], and the transportation costs, estimated by the time spent on transportation and standard mileage reimbursement [32]. Table 3 presents detailed cost specification of BHF program online training and delivery.

Table 3 Labor activity specification of implementing BHF intervention in the HEI trial by the Proctor framework [24]

Non-labor costs

Non-labor costs include gym and classroom space, scales, diet books, printouts, handouts, educational supplies, equipment, and financial incentives provided for BHF participants and will be collected based on actual amounts spent and tracked from receipts and paid invoices. Costs of computers, Internet, and landline will not be included given that they are provided to community facilitators to conduct their daily tasks. To supplement the cost estimate of D&I strategies, we will also obtain cost information during the exit interview of community facilitators, as suggested by the literature [14].

Data collection

To track all the activities associated with a D&I strategy and BHF program online training, recruitment and delivery, we use a publicly available time-tracking tool (i.e., Clockify), which is accessible through the website or smartphone app, to monitor the activities logging in by the community facilitators or the research team. To mitigate the potential respondent burden, we pre-populate the identified implementation activities associated with each D&I strategy proposed in the trial on the Clockify platform. Community facilitators in both conditions will be asked to record their BHF associated activities if any (typically the day before and the day of the BHF sessions) using either the web- or mobile-based Clockify app. They will receive weekly reminders for weekly implementation activity tracking, regardless of their group assignment, through an automated system email sent out during the weekend. In addition, the project coordinator will engage with the communities for reminders via emails, text messages, or phone calls if any given data was not logged in the system. We will supplement the cost data collection with meeting attendance sheets and minutes (for the research team and CAB members) and a BHF session check-in sheet (for participants).

Data analysis

Table 1 summarizes the proposed within-trial economic evaluation and associated costs and outcomes by the RE-AIM indicators. The cost-effectiveness of D&I strategies is assessed by combing their associated costs and outcomes using an incremental implementation cost-effectiveness ratio [33] (ICER) if a strategy is simply applied to the BHF-LC with the BHF-PO as the comparison group or average implementation cost-effectiveness ratio [34] (ACER) if a D&I strategy is applied in both the BHF-LC and BHF-PO conditions (and thus no comparison group). The interpretation is similar to the interpretation for a trationdtional ICER. An ICER is interpreted as the extra cost incurred for the intervention to achieve an extra 1 unit of increase in outcomes (reach, adoption, implementation, and maintenance) measured, whereas ACER represents the cost associated with a one-unit change on given measures for the outcome of interest. For example, the ACER for the fund and contract dissemination strategy to increase the adoption of the BHF program, at the micropolitan and surrounding rural areas is described as the costs associated with one community responding to RFA and contracted to implement BHF.

Similarly, the cost-effectiveness of the BHF program is determined by the total LC costs (BHF-LC group only) and the BHF program costs consisting of training, recruitment, program delivery, and post-follow-up, and the effect of BMI z-score reduction at 6 and 12 months. The ICER result is interpreted as the cost for an additional unit decrease in BMI z-scores between the BHF-LC and BHF-PO conditions.

To ascertain whether a given ICER/ACER is cost-effective, we will compare them to thresholds that are based on findings from previous economic research [13].

Sensitivity analysis

Performing a sensitivity analysis is a critical component of an economic evaluation when the assumptions and parameter estimates applied in the analysis have uncertain precision. Two approaches will be employed in the proposed economic evaluation. The first approach is one-way deterministic sensitivity analysis (DSA) [35] for the cost analysis of D&I strategies and BHF program training, recruitment and delivery. We will conduct the DSA around key unit costs, cost components, and outcome measures (i.e., BMI z-score) to assess the impact of these changes/variations on the total costs and cost-effectiveness [36, 37]. We will use tornado diagrams to summarize the effects of varying key input parameters one at a time on the results of the ICER/ACER [38]. Parameters to be considered include but are not limited to the role and salary level of key personnel, intensity/number of BHF sessions, and changes in the lengths of attendance time for the session. The parameters will be sorted in descending order by their influence on the ICER/ACER. The longer bars indicate the most important parameters. The second approach is a nonparametric bootstrap simulation, and it will be used to construct the joint uncertainty (e.g., 95% confidence interval) for the incremental costs and incremental effects (i.e., BMI z-score reduction) between BHF-LC and BHF-PO conditions. In the bootstrap simulation, 2000 random samples of cost-effect pairs will be selected with replacement from the original trial dataset. Simulation results will be presented graphically using a scatter plot (i.e., the incremental cost-effectiveness plane [ICEP]) [39], where each dot represents the ICER of one iteration of the bootstrap simulation. In addition, we will conduct CEA of BHF intervention using parents’ weight loss outcomes at 6 and 12 months. Moreover, we will conduct scenario analyses under different settings and conditions to estimate replication costs to enhance decision-making for potential adopting communities [40, 41].

All analyses will be performed using Stata 16.0 (StataCorp LP, College Station, TX) or Microsoft Excel.

Discussion

This protocol sets out our plans to assess the cost and cost-effectiveness of D&I strategies, utilized in a type III HEI trial to promote the adoption of an evidence-based PWMI (i.e., BHF) via a trial-based economic evaluation. Through this publication, we aim to inform the research and public health communities of the potential cost and benefits of implementing this type of work. Moreover, we will provide a case study of an economic plan by the RE-AIM indicators.

Economic consideration is often critical to decision makers who have to decide about the cost of adopting, implementing, and sustaining evidence-based interventions (EBIs) [42] as well as the costs related to D&I strategies applied to increase the rate of adoption, quality of implementation, and the likelihood of sustainability. Powell and colleagues [17] identified the need to increase economic evaluations of D&I strategies as one of five research priorities for advancing D&I science. Current areas of challenge in this pursuit include a lack of distinction between strategies that would promote adoption of an EBI with strategies that would promote high-quality implementation, and strategies that would improve the likelihood of sustainability, as well as the lack of a pragmatic and generalizable approach for gathering information on strategy costs, capturing adaptations and changes in strategies made during implementation, and the conduct of implementation economic evaluation.

These stem from a fundamental difference between implementation economic evaluation and traditional economic evaluation. The former addresses the resources (e.g., implementation activities) needed to encourage adoption, implementation, and sustainability of the EBI, whereas the latter focuses primarily on the start-up and ongoing costs of the EBI itself [43]. D&I strategies can vary in their intensity and resource use as well as their effectiveness in aiding the implementation of EBIs [6]. In this economic evaluation protocol targeting D&I strategies, we attempt to identify and present the strategy costs across RE-AIM dimensions to clarify sometimes fuzzy boundaries between these strategies. Moreover, to conduct cost assessment across RE-AIM dimensions, it may be implicitly assumed that one implementation strategy has an impact on only one implementation outcome, which is likely not the case in the present study. For example, the strategy of “developing packaged implementation blueprint” has impacts on three implementation outcomes (reach, adoption, and implementation). Even a fund and contract dissemination strategy with a primary goal to increase adoption embedded the need for network weaving to improve implementation and likely sustainability over the life course of the program. To not overestimate the cost-effectiveness (i.e., double-counting) of a D&I strategy, we deliberately attributed the activities within the strategy to an associated primary implementation outcome target for the given strategy.

Furthermore, we conduct CEA using costs related to each implementation outcome to demonstrate the potential to separate the strategy costs and present the CEA results by RE-AIM outcomes. Most importantly, we follow labels and definitions for the implementation strategy components based on the Expert Recommendations for Implementing Change (ERIC) taxonomies [6], which is commenced to address the challenges and difficulties of reconciling the terms and definitions of strategies, to allow for comparisons of our work to other studies examining the utility of different implementation strategies. However, we also recognize the challenge that the application of ERIC taxonomies is less than optimal in the real-world settings, especially in the community context—no clear cut between the discrete ERIC strategies and also the combinations of ERIC strategies and non-ERIC strategies. Moreover, the strategies used may change over the life course of intervention. It is critical to systematically document the adaptation made during the implementation of the programs [44], using the well-known Stirman framework [45], the LISTS approach proposed by Smith et al. [46, 47], or the FRAME-IS approach [48], and record the costs associated with adaptation activities along the course of the intervention.

Identification of costs incurred when adopting, implementing, and delivering an EBI is key to addressing the dissemination and implementation of EBIs and the strategies used to close the research-practice gap [49,50,51,52]. Unfortunately, costs are infrequently collected or reported with limited quality (mainly due to a non-standardized approach) and scope for understanding the costs or cost-effectiveness of implementation strategies [53,54,55]. There is an urgent need to provide a pragmatic and replicable methodology for costing D&I strategies to help stakeholders or community organizations decide what strategies are most efficient at achieving the goal of adoption, implementation, and sustainability. To facilitate the cost reporting of D&I strategies, several efforts have been made to the costing methods and applications [14, 30, 56,57,58]. In two recent studies, for example, Jordan et al. [56] reported using electronic cost capture methods to track personnel time associated with implementation preparation activities of a pediatric obesity intervention in primary care. Ritchie et al. [57] estimated the time and organization costs of facilitating the implementation of mental health integration in primary care with a structured spreadsheet including pre-specified facilitation activities, duration, and participating personnel. Although their approach may provide an accounting of resources for a particular setting, these results may not accurately generalize to other settings or be comparable since they simply reported partial costs in a subset of activities rather than across the entire phases of the implementation (pre-implementation costs in Jordan et al. [56] and implementation costs in Ritchie et al. [57]).

Notably, Cidav and colleagues [30] presented a pragmatic costing method integrating TDABC and the Proctor framework for the identification, specification, and reporting of implementation strategies [24]. However, the three parameters (frequency of the activity, average time to perform the activity, and unit costs of the resources to perform the activity) included in the TDABC may not appropriately capture costs or apply to implementation strategies due to intertwined elements and activities incorporated in each discrete D&I strategy. Furthermore, they did not account for the opportunity costs incurred by the action target (individuals who received the intervention) in the cost estimate formula of implementation strategies. Capturing implementation strategies can be challenging, as responsibility for implementation is often diffuse and strategies may be flexibly applied as barriers and challenges emerge [59]. Our proposed pragmatic and generalizable costing approach for D&I strategies with modified activity-based costing integrated with the Proctor framework applying a publicly available software will facilitate the cost estimate and tracking ability of all participating communities. The results will also demonstrate the usability of the proposed time-tracking tool for implementation activities to facilitate cost assessments of D&I strategies in the community settings, which holds the potential to be further disseminated to community-based practice/organizations with limited capacity or resources for cos tracking. Furthermore, the replication costs derived from the scenario analysis may shed some light on addressing the economic consideration of research translation by providing informed decision making for stakeholders who are interested in adopting the program with the aforementioned D&I strategies in their communities [41].

Challenges and limitations

To our knowledge, this is one of few studies to conduct the economic evaluation by distinguishing among strategies that would promote adoption of an EBI from those that would promote high-quality implementation and from those that would improve the likelihood of sustainability. While we are setting out the venue to distinguish and collect the data to establish the economic case of implementation and D&I strategies in a type III HEI trial, there are a number of anticipated challenges of the work and associated potential limitations. First, it is of great possibility that community facilitators may not constantly record their implementation activities regardless of the ease of use of digital-based time tracking tools and the regular reminders sent out by the research team and the time tracking system. To remedy this concern, the documented activities will be complemented by the information gathered through research team field notes and from exit interviews, and also the activity log tracked by the research team for a robust cost assessment of D&I strategies. Second, the HEI trial is being conducted in the micropolitan and surrounding rural communities of Nebraska, where system and population characteristics may vary across different counties/communities. The proposed costing approach and economic evaluation framework may not be applicable to other non-micropolitan settings. To improve the replicability, we will explore the potential of the proposed economic evaluation in other settings, populations, and implementation strategies in an attempt to detangle the complexity of the work as described above and also the results under replication costs. The goal of this pilot economic evaluation is focusing more scientifically on teasing out costs by specific D&I strategies related to specific D&I outcomes. As such, our study does not address the use of these cost evaluations for decision makers, though we hypothesize that by engaging our CAB and community implementation teams on reviewing our findings that we will gain insight on how better to communicate cost evaluation as a decision aid in communities considering PWMI implementation. Finally, due to the limited sample size of the study population of this protocol (n=21–28 community facilitators), it may be underpowered to detect the cost-effectiveness of a LC implementation strategy plus BHF Online Training Resources and Program Package (BHF-LC) compared to BHF Online Training Resources and Program Package access only (BHF-PO) on the adoption, implementation, and sustainability of implementing the BHF program.

Conclusion

The proposed costing approach and results of the economic evaluation of D&I strategies and implementing an EBI will contribute to the evolving but still scant literature on costing implementation strategies (and especially those for adoption and sustainment) and fill the evidence gap on the cost-effectiveness of D&I strategies. The results should assist policy makers, community organizations, local health departments, and stakeholders in potential adopting settings for an informed decision making regarding the sustainability and future adoption and dissemination.

Availability of data and materials

There are no datasets associated with this protocol.

Abbreviations

EBIs:

Evidence-based interventions

BHF:

Building Health Families

CEA:

Cost-effectiveness analysis

HEI:

Hybrid effectiveness-implementation trial

LC:

Learning collaborative

RE-AIM:

Reach, effectiveness, adoption, implementation, and maintenance

PWMIs:

Pediatric weight management interventions

D&I:

Dissemination and implementation

CAB:

Community advisory board

RFA:

Request for applications

TLD:

Traffic light diet

BMI:

Body mass index

TDABC:

Time-driven activity-based costing

LOI:

Letter of intent

ICER:

Incremental implementation cost-effectiveness ratio

ACER:

Average implementation cost-effectiveness ratio

DSA:

Deterministic sensitivity analysis

ICEP:

Incremental cost-effectiveness plane

ERIC:

Expert Recommendations for Implementing Change

FRAME-IS:

Framework for Reporting Adaptations and Modifications to Evidence-based Implementation Strategies

LISTS:

Longitudinal Implementation Strategies Tracking System

References

  1. Weihrauch-Blüher S, Schwarz P, Klusmann J-H. Childhood obesity: increased risk for cardiometabolic disease and cancer in adulthood. Metabolism. 2019;92:147–52.

    Article  PubMed  Google Scholar 

  2. Llewellyn A, et al. Childhood obesity as a predictor of morbidity in adulthood: a systematic review and meta-analysis. Obes Rev. 2016;17(1):56–67.

    Article  CAS  PubMed  Google Scholar 

  3. Hill JL, et al. A type III hybrid effectiveness–implementation pilot trial testing dissemination and implementation strategies for a pediatric weight management intervention: the Nebraska Childhood Obesity Research Demonstration Project. Child Obes. 2021;17(Supplement 1):S-70–8.

    Article  Google Scholar 

  4. Estabrooks PA, et al. Using integrated research-practice partnerships to move evidence-based principles into practice. Exerc Sport Sci Rev. 2019;47(3):176–87.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Heelan KA, et al. Packaging of a pediatric weight management intervention and implementation blueprint for rural and micropolitan communities: the Nebraska CORD 3.0 Project. Child Obes. 2021;17(Supplement 1):S-62–9.

    Article  Google Scholar 

  6. Powell BJ, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):21.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Glasgow RE, et al. RE-AIM Planning and Evaluation Framework: adapting to new science and practice with a 20-year review. Front Public Health. 2019;7:64.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Schouten LM, et al. Evidence for the impact of quality improvement collaboratives: systematic review. BMJ. 2008;336(7659):1491–4.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Nadeem E, et al. Understanding the components of quality improvement collaboratives: a systematic literature review. Milbank Q. 2013;91(2):354–94.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Dopp AR, et al. Translating economic evaluations into financing strategies for implementing evidence-based practices. Implement Sci. 2021;16(1):66.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Eisman AB, et al. Implementation science issues in understanding, collecting, and using cost estimates: a multi-stakeholder perspective. Implement Sci. 2021;16(1):75.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Barnett ML, et al. Moving beyond Aim Three: a need for a transdisciplinary approach to build capacity for economic evaluations in implementation science. Implement Sci Commun. 2021;2(1):133.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Dopp AR, et al. Community-based implementation of trauma-focused interventions for youth: economic impact of the learning collaborative model. Psychol Serv. 2017;14(1):57.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Dopp AR, et al. Mixed-method approaches to strengthen economic evaluations in implementation research. Implement Sci. 2019;14(1):2.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Roberts SLE, Healey A, Sevdalis N. Use of health economic evaluation in the implementation and improvement science fields—a systematic literature review. Implement Sci. 2019;14(1):72.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Reeves P, et al. Economic evaluations of public health implementation-interventions: a systematic review and guideline for practice. Public Health. 2019;169:101–13.

    Article  CAS  PubMed  Google Scholar 

  17. Powell BJ, et al. Enhancing the impact of implementation strategies in healthcare: a research agenda. Front Public Health. 2019;7:3.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Gold HT, et al. Cost data in implementation science: categories and approaches to costing. Implement Sci. 2022;17(1):11.

    Article  PubMed  PubMed Central  Google Scholar 

  19. National Cancer Institute. Building capacity for economic evaluation in implementation science. 2021; Available from: https://cancercontrol.cancer.gov/is/blog/building-capacity-for-economic-evaluation. Cited 2022 January 18.

    Google Scholar 

  20. Ruebel ML, et al. Outcomes of a family based pediatric obesity program-preliminary results. Int J Exerc Sci. 2011;4(4):217.

    PubMed  PubMed Central  Google Scholar 

  21. Heelan KA, et al. Healthier school environment leads to decreases in childhood obesity: the Kearney Nebraska story. Child Obes. 2015;11(5):600–7.

    Article  PubMed  Google Scholar 

  22. Abbey B, Heelan K, Bartee T. Overcoming barriers to a school BMI screening program. In: Annals of behavioral medicine. Cary: Oxford Univ Press Inc Journals Dept; 2017.

    Google Scholar 

  23. Golden CA, Hill JL, Heelan KA, Bartee RT, Abbey BM, Malmkar A, et al. A dissemination strategy to identify communities ready to implement a pediatric weight management intervention in medically underserved areas. Prev Chronic Dis. 2021;18:200248. https://doi.org/10.5888/pcd18.200248.

  24. Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. 2013;8(1):139.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Rogers EM. Diffusion of innovations. Simon and Schuster. New York; 2003.

  26. Chambers DA, Norton WE. The adaptome: advancing the science of intervention adaptation. Am J Prev Med. 2016;51(4):S124–31.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Edmunds JM, Beidas RS, Kendall PC. Dissemination and implementation of evidence–based practices: training and consultation as implementation strategies. Clin Psychol Sci Pract. 2013;20(2):152.

    Article  Google Scholar 

  28. Husereau D, et al. Consolidated health economic evaluation reporting standards (CHEERS) statement. Cost Eff Resour Alloc. 2013;11(1):6.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Jones Rhodes WC, Ritzwoller DP, Glasgow RE. Stakeholder perspectives on costs and resource expenditures: tools for addressing economic issues most relevant to patients, providers, and clinics. Transl Behav Med. 2018;8(5):675–82.

    Article  PubMed  Google Scholar 

  30. Cidav Z, et al. A pragmatic method for costing implementation strategies using time-driven activity-based costing. Implement Sci. 2020;15(1):1–15.

    Article  Google Scholar 

  31. U.S. Bureau of Labor Statistics- Occupational Employment and Wage Statistics. Available from: https://www.bls.gov/oes/current/oes_nat.htm#00-0000. Cited 2021 July 9.

  32. U.S. Internal Revenues Service- Standard Mileage Rates. Available from: https://www.irs.gov/tax-professionals/standard-mileage-rates. Cited 2021 December 7.

  33. Raghavan R. The role of economic evaluation in dissemination and implementation research. In: Dissemination and implementation research in health: translating science to practice. New York: Oxford University Press; 2018. p. 89–106.

    Google Scholar 

  34. Briggs A, Fenn P. Trying to do better than average: a commentary on ‘statistical inference for cost-effectiveness ratios’. Health Econ. 1997;6(5):491–5.

    Article  CAS  PubMed  Google Scholar 

  35. Jain R, Grabner M, Onukwugha E. Sensitivity analysis in cost-effectiveness studies. Pharmacoeconomics. 2011;29(4):297–314.

    Article  PubMed  Google Scholar 

  36. Michaud TL, et al. Cost analysis of a remote patient monitoring programme for post-discharge patients with type 2 diabetes. J Telemed Telecare. 2021;0(0):1357633X20985393.

    Google Scholar 

  37. Michaud TL, et al. Costing a population health management approach for participant recruitment to a diabetes prevention study. Transl Behav Med. 2021;11(10):1864–74.

    Article  PubMed  PubMed Central  Google Scholar 

  38. Groot Koerkamp B, et al. Uncertainty and patient heterogeneity in medical decision models. Med Decis Mak. 2010;30(2):194–205.

    Article  Google Scholar 

  39. Briggs AH, O'Brien BJ, Blackhouse G. Thinking outside the box: recent advances in the analysis and presentation of uncertainty in cost-effectiveness studies. Annu Rev Public Health. 2002;23(1):377–401.

    Article  PubMed  Google Scholar 

  40. Ritzwoller DP, et al. Economic analyses of the Be Fit Be Well program: a weight loss program for community health centers. J Gen Intern Med. 2013;28(12):1581–8.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Ritzwoller DP, et al. Intervention costs and cost-effectiveness for a multiple-risk-factor diabetes self-management trial for Latinas: economic analysis of¡ Viva Bien! Transl Behav Med. 2011;1(3):427–35.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Neta G, et al. A framework for enhancing the value of research for dissemination and implementation. Am J Public Health. 2015;105(1):49–57.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Eisman AB, et al. Economic evaluation in implementation science: making the business case for implementation strategies. Psychiatry Res. 2020;283:112433.

  44. Patient-Centered Outcomes Research Institute. PCORI methodology standards. 2019; Available from: https://www.pcori.org/sites/default/files/PCORI-Methodology-Standards.pdf. Cited 2022 February 1.

    Google Scholar 

  45. Stirman SW, et al. Development of a framework and coding system for modifications and adaptations of evidence-based interventions. Implement Sci. 2013;8(1):1–12.

    Article  Google Scholar 

  46. Smith JD, Li DH, Rafferty MR. The implementation research logic model: a method for planning, executing, reporting, and synthesizing implementation projects. Implement Sci. 2020;15(1):1–12.

    Article  Google Scholar 

  47. Smith J, et al. A longitudinal implementation strategies tracking system: development and initial acceptability. In: 13th annual virtual conference on the science of dissemination and implementation: AcademyHealth; 2020.

  48. Miller CJ, et al. The FRAME-IS: a framework for documenting modifications to implementation strategies in healthcare. Implement Sci. 2021;16(1):36.

    Article  PubMed  PubMed Central  Google Scholar 

  49. Hoomans T, Severens JL. Economic evaluation of implementation strategies in health care. Implement Sci. 2014;9(1):1–6.

  50. Aarons GA, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Adm Policy Ment Health Ment Health Serv Res. 2011;38(1):4–23.

    Article  Google Scholar 

  51. Saldana L, et al. The cost of implementing new strategies (COINS): a method for mapping implementation resources using the stages of implementation completion. Child Youth Serv Rev. 2014;39:177–82.

    Article  PubMed  Google Scholar 

  52. Glasgow RE, Emmons KM. How can we increase translation of research into practice? Types of evidence needed. Annu Rev Public Health. 2007;28:413–33.

    Article  PubMed  Google Scholar 

  53. Hoomans T, et al. The methodological quality of economic evaluations of guideline implementation into clinical practice: a systematic review of empiric studies. Value Health. 2007;10(4):305–16.

    Article  PubMed  Google Scholar 

  54. Grimshaw, J., et al., Effectiveness and efficiency of guideline dissemination and implementation strategies. 2004.

    Book  Google Scholar 

  55. Vale L, et al. Systematic review of economic evaluations and cost analyses of guideline implementation strategies. Eur J Health Econ. 2007;8(2):111–21.

    Article  PubMed  Google Scholar 

  56. Jordan N, et al. Costs of preparing to implement a family-based intervention to prevent pediatric obesity in primary care: a budget impact analysis. Prev Sci. 2019;20(5):655–64.

    Article  PubMed  PubMed Central  Google Scholar 

  57. Ritchie MJ, et al. Time and organizational cost for facilitating implementation of primary care mental health integration. J Gen Intern Med. 2020;35(4):1001–10.

    Article  PubMed  Google Scholar 

  58. Lang JM, Connell CM. Measuring costs to community-based agencies for implementation of an evidence-based practice. J Behav Health Serv Res. 2017;44(1):122–34.

    Article  PubMed  Google Scholar 

  59. Bunger AC, et al. Tracking implementation strategies: a description of a practical approach and early findings. Health Res Policy Syst. 2017;15(1):15.

    Article  PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

This study was supported by the Centers for Disease Control and Prevention (CDC) (Award No: 1U18DP006431-01-00). This work is solely the responsibility of the authors and does not represent official views of the Centers for Disease Control and Prevention (CDC), Health and Human Services (HHS), or the US Government.

Author information

Authors and Affiliations

Authors

Contributions

Study conceptualization: TLM, PAE, JLH, KAH, and REG. Writing: TLM, PAE, and REG. Editing: TLM, PAE, JLH, KAH, REG, RTB, BMA, AM, JM, CG, and GP. The authors read and approved the final manuscript.

Corresponding author

Correspondence to Tzeyu L. Michaud.

Ethics declarations

Ethics approval and consent to participate

The economic evaluation includes a within-trial cost-effectiveness analysis. The trial protocol which included the collection of cost data used in this study has been approved by the University of Nebraska at Kearney Institutional Review Board. Results of the main trial and economic evaluation will be presented in academic conferences and published in the peer-reviewed journals.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

CHEERS Checklist.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Michaud, T.L., Hill, J.L., Heelan, K.A. et al. Understanding implementation costs of a pediatric weight management intervention: an economic evaluation protocol. Implement Sci Commun 3, 37 (2022). https://doi.org/10.1186/s43058-022-00287-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s43058-022-00287-1

Keywords