Skip to main content

A picture is worth a thousand words: advancing the use of visualization tools in implementation science through process mapping and matrix heat mapping

Abstract

Background

Identifying key determinants is crucial for improving program implementation and achieving long-term sustainment within healthcare organizations. Organizational-level complexity and heterogeneity across multiple stakeholders can complicate our understanding of program implementation. We describe two data visualization methods used to operationalize implementation success and to consolidate and select implementation factors for further analysis.

Methods

We used a combination of process mapping and matrix heat mapping to systematically synthesize and visualize qualitative data from 66 stakeholder interviews across nine healthcare organizations, to characterize universal tumor screening programs of all newly diagnosed colorectal and endometrial cancers and understand the influence of contextual factors on implementation. We constructed visual representations of protocols to compare processes and score process optimization components. We also used color-coded matrices to systematically code, summarize, and consolidate contextual data using factors from the Consolidated Framework for Implementation Research (CFIR). Combined scores were visualized in a final data matrix heat map.

Results

Nineteen process maps were created to visually represent each protocol. Process maps identified the following gaps and inefficiencies: inconsistent execution of the protocol, no routine reflex testing, inconsistent referrals after a positive screen, no evidence of data tracking, and a lack of quality assurance measures. These barriers in patient care helped us define five process optimization components and used these to quantify program optimization on a scale from 0 (no program) to 5 (optimized), representing the degree to which a program is implemented and optimally maintained. Combined scores within the final data matrix heat map revealed patterns of contextual factors across optimized programs, non-optimized programs, and organizations with no program.

Conclusions

Process mapping provided an efficient method to visually compare processes including patient flow, provider interactions, and process gaps and inefficiencies across sites, thereby measuring implementation success via optimization scores. Matrix heat mapping proved useful for data visualization and consolidation, resulting in a summary matrix for cross-site comparisons and selection of relevant CFIR factors. Combining these tools enabled a systematic and transparent approach to understanding complex organizational heterogeneity prior to formal coincidence analysis, introducing a stepwise approach to data consolidation and factor selection.

Peer Review reports

Background

Identification of common patterns across organizations can improve understanding of complex, interactive factors that facilitate and inhibit implementation processes and outcomes. However, studying the implementation and long-term sustainment of complex interventions involving multidisciplinary teams across organizations can be challenging. Examples of challenges include summarizing vast amounts of data from multiple stakeholders, reconciling inconsistencies, and identifying patterns across programs [1, 2]. Data visualization methodologies can assist with analysis and data presentation when comparing complex systems across organizations. Data visualization methods such as service blueprinting, journey mapping, and ecomapping have all been used in service design and health services research [3,4,5]. However, these methods focus primarily on “end-user” perspectives and illustrating care networks rather than highlighting organizational processes.

Process mapping, heat mapping, and data matrices are data visualization techniques that can help to summarize complex processes or multi-level data. A process map is a diagram or flow chart representing a sequence of actions to assist stakeholders in visualizing a given process or workflow [6]. Process mapping has been previously applied in health services research [7], quality improvement (QI), and quality assurance (QA) initiatives, but has not been widely used for contrasting processes/procedures across multiple sites in implementation research. Heat mapping is a data visualization technique that uses color to show magnitude differences within multiple conceptual ideas or data points. Heat maps are widely used in bioinformatics to visualize large gene expression data sets but have been underutilized in other areas [8, 9], including implementation science. Lastly, data matrices have been used to organize data and analyze complexities across multiple stakeholders and organizations. Constructing data matrices using the “Framework method” [10] helps summarize qualitative data from each stakeholder in a separate row of a spreadsheet and organizes data by themes within columns. The “matrixed multiple case study approach” recently described by Kim et al. [2] documents and compares multiple data sources both within and across organizational units to understand factors associated with program implementation. This method helped organize complex data into sortable matrices for cross-case comparisons, and researchers designated a single column to assign values for impact factors on program implementation. In their study, values were defined as “enabling,” “hindering,” or “neutral,” and success was defined based on implementation outcomes from Proctor et al. [11] Another example of using data matrices in implementation science is the “rapid-cycle evaluation approach to improving implementation” [12] where matrices helped organize data and support a cross-case analysis of patterns with barriers and facilitators. Each of these three approaches has been used to organize and analyze complex data. Although these have been widely used in other types of research, their use in implementation science is limited. We describe how we combined and built upon these methods with color-coding to aid in both data visualization and further consolidation of data, a method we call data matrix heat mapping.

We provide an example to demonstrate how process mapping and data matrix heat mapping can be used together to facilitate cross-case comparisons of factors that may impact implementation of a complex intervention or program at the organizational level. As part of the Implementing Universal Lynch Syndrome screening study (IMPULSS) [13], we refined and applied these methods to compare intervention protocols and processes across sites and identify factors relevant for program implementation and optimization. This methods manuscript details how the novel combination of process mapping and matrix heat mapping helped achieve our study objectives. The step-by-step methods we describe could be applied and/or modified for use in other implementation science research.

Methods

Study setting

The goal of IMPULSS is to characterize variation in healthcare system processes used to screen individuals with endometrial and colorectal tumors to identify Lynch syndrome (LS), a hereditary cancer predisposition condition [14]. This screening process, commonly referred to as universal tumor screening (UTS), is a multi-step complex intervention that requires coordination across multiple departments and stakeholders within a health system. Ultimately, the goal of a UTS program is to identify more patients with LS who will benefit from condition-specific medical care by identifying and/or preventing cancers early, thereby improving patient outcomes. UTS for both colorectal and endometrial cancers has been found to be both feasible and beneficial [15, 16], with support from several professional organizations including the National Comprehensive Cancer Network (NCCN), the American College of Obstetrics and Gynecology (ACOG), the American Society for Clinical Pathology (ASCP), and the Evaluation of Genomic Applications in Practice and Prevention (EGAPP) working group [17,18,19,20,21]. Despite strong evidence of patient benefit and over a decade of support for UTS, a standard approach for implementing a program is lacking, and variability has been identified within UTS protocols across healthcare organizations [22, 23]. Hence, key IMPULSS study objectives were to compare variation in site-specific protocols, determine contextual factors that might make a difference in organizational decision-making to implement UTS programs, and identify components related to program optimization.

The IMPULSS study design is described in detail elsewhere [13]. Briefly, nine healthcare systems participating in IMPULSS were chosen for their variability in having or not having a UTS program. To understand program variation and barriers and facilitators to UTS program existence or implementation, data collection consisted of four researchers interviewing 66 stakeholders across the healthcare systems. A semi-structured interview guide was developed and employed based on the Consolidated Framework for Implementation Research (CFIR) and contained several additional questions to elucidate processes of the tumor screening protocols at each organization (see Fig. 1 of supplemental materials for example). Interviews were conducted by ZS, IL, VS, and JA by telephone or Microsoft Teams, were recorded, transcribed verbatim, and coded as outlined below in step 1 of both process mapping and data matrix heat mapping. CFIR constructs, combined constructs, and domains [24] are hereafter referred to as “factors” to simplify terminology. The process mapping and matrix heat mapping methods described below are all constructed from interview data.

Process mapping methodology and output

The goal of process mapping was to (1) visually represent and easily compare each organization’s UTS protocol, (2) help visually pinpoint barriers to patient care that may affect process optimization, and (3) identify components to help quantify implementation optimization levels. Process mapping was led by ZS, with consultation from study team members, and guided by literature [7], including a previous study conducting cross-case comparison using data visualization of UTS protocols [25] and a study looking at factors associated with patient follow-through after a positive screen [22]. We used coded interviews combined with prior literature [22, 25] to conduct process mapping. Our approach to process mapping was highly iterative and consisted of six main steps (see Fig. 1) outlined below.

Fig. 1
figure 1

Process mapping methodology to quantify level of implementation success (i.e., optimization scores)

Step 1: code protocol descriptions from each stakeholder using a data extraction sheet to identify process steps

First, ZS and JA, along with a research assistant, reviewed two stakeholder transcripts to create an initial data extraction sheet of key components and procedures (i.e., process steps) that make up UTS protocols informed by prior research [22, 25] and NCCN guidelines [17]. All stakeholder transcripts from a single organization were combined and analyzed together. Discussions among authors resulted in clarifications to the extraction sheet as questions arose while completing data extraction and coding of the remaining transcripts. During this process, commonalities and differences in procedures were identified across and within organizations.

Step 2: list process and contextual differences to define organizational units

During step 1, conflicting information between some stakeholders within the same healthcare organization prompted us to list process inconsistences and contextual differences in both structural characteristics and geographic location. ZS, AKR, and DC listed and sorted stakeholders by these differences (see Fig. 2 of supplemental materials for example), occasionally identifying different protocols within a single healthcare organization. Organizational units were defined as unique sites for analysis, each with their own protocol for identifying patients with LS.

Step 3: create visual representations of protocols to document processes and stakeholder inconsistencies within each organizational unit

ZS reviewed data from Steps 1 and 2 to create a process map of each organization’s approach to identifying LS (see Fig. 3 of supplemental materials for example). These preliminary process maps highlighted discrepancies across stakeholders and helped lead to the creation of additional process maps if multiple organizational units were identified during the prior step or current step.

Step 4: reconcile stakeholder inconsistencies to validate and finalize the process map for each organizational unit

ZS and DC conducted formal reconciliation and validation of the process map for each organizational unit. Minor discrepancies that remained after creating separate maps were resolved by weighing perspectives differently depending on the roles and responsibilities of the interviewed participant. Greater emphasis was given to the perspectives of primary stakeholders, defined as a provider stakeholder working directly with implementation and/or maintenance within the UTS protocol. ZS and DC led member checking where process maps were shown to respective stakeholders. Member check meetings were used to share, confirm, reconcile, and clarify collected data and related interpretations among stakeholders. Feedback was used to adjust process maps as needed. The outcome of this step was a final process map for each of the 19 organizational units (see Fig. 4 of supplemental materials for example).

Step 5: compare process maps across organizational units to find gaps and inefficiencies and define process optimization components

Once each process map was finalized, we sought to find commonalities in process optimization components between organizational units. ZS, in consultation with AKR and DC, conducted a cross-case comparison by cross-referencing each process map with coded transcripts and highlighting similarities and differences between organizational units. This comparison via data visualization helped ZS, AKR, and DC find common points within the UTS protocols where tumor screening and/or genetics referrals were not completed. These patient care barriers were listed and defined as process gaps and inefficiencies that were used for optimization scoring in the next step (see Fig. 5 of supplemental materials).

Step 6: create overall scores representing optimization levels for each organizational unit

ZS, DC, and AKR settled on five components associated with an optimized UTS program to help create an optimization scoring system. These components were selected based on descriptions by interviewed participants when discussing process execution and program evaluation and informed by previous findings [22, 25]. Each organizational unit was coded as “1”, “0.5”, or “0” for each optimization component using the following definitions determined by the study team: A “1” indicated presence of the component, a “0” indicated absence, and “0.5” was typically assigned if the organizational unit previously had an optimization component, such as quality assurance, that they no longer conduct or is no longer present. Additionally, one organizational unit received a “0.5” because it remained unresolved due to stakeholder discrepancy. Upon totaling the optimization scale, sites that had implemented a UTS protocol could receive a maximum total score of 5 and those that had not implemented a protocol were assigned a total score of 0. During our member-checking process, we shared the optimization scores to verify their accuracy and “face validity” according to stakeholder experience. We incorporate these scores in a subsequent step within the data matrix heat mapping process described below (step 5).

Data matrix heat mapping methodology and output

Steps for data matrix heat mapping were developed concurrently as the study team operationalized the outcome using process mapping. The goal of matrix heat mapping was to consolidate, organize, and select relevant CFIR factors as a necessary, preparatory step for our planned future analysis. ZS and AKR have prior experience with using the Framework method [10] to organize interview data, and DC has prior experience using color-coding as a visual method to designate differences across cases. Additional inspiration for our approach came from the multiple matrixed case study approach [2] and prior studies that applied CFIR coding to qualitative data [24, 26, 27]. Data matrix heat mapping (illustrated in Figs. 713 of the supplementary materials) consisted of seven main steps (see Fig. 2) to compile and code contextual data within each organizational unit using multiple matrices that were ultimately combined into a single data matrix to compare contextual factors and outcomes across all units.

Fig. 2
figure 2

Data matrix heat mapping methodology to compile, organize, and consolidate complex data

At various points throughout the data matrix heat mapping methodology, ZS and DC referred to transcripts and conducted member-checking with primary stakeholders and site PIs (AKR, JLS, MME, CYL, PAP, RNS, SYL, ANBH, and JEH) for verification and to evaluate the context in which quotations or summaries were extracted [28]. This type of iterative approach was applied to capture nuance within the rich data and to improve the rigor and accuracy of coding.

Step 1: conduct formal qualitative analysis on stakeholder data

A formal qualitative coding process guided by the CFIR framework was employed with interview transcripts by a two-person team of JBA and DC. First, JBA and DC used the CFIR codebook to code two interview transcripts. Throughout early coding, JBA and DC made modifications to the codebook with consultation from the study team to add inclusion and exclusion criteria specific for UTS implementation. Next, JBA continued coding all transcripts using RQDA qualitative software [29], with DC sorting codes by CFIR factor and reviewing text coded within each factor. Throughout coding, discrepancies were discussed and resolved by JBA and DC with input from additional study personnel as needed.

Step 2: create a data matrix for each organizational unit to compile and organize stakeholder data

DC and two research assistants summarized the coded data to create a data matrix for each organizational unit. To do this, CFIR factors were listed in columns and relevant quotes/summaries were included in the respective cells such that responses from each stakeholder interviewed were represented across a single row within the data matrix (i.e., spreadsheet) for their respective organizational unit.

Step 3: assign valences using color-coding to begin heat mapping

We applied a modified CFIR data coding approach to assigning valences that included the use of color-coding [26, 30]. Specifically, ZS and DC reviewed each data matrix and independently color-coded the summaries for each construct for each stakeholder according to the following valences: positive or presence of facilitator = blue, negative or presence of a barrier = red, mixed (i.e., both positive and negative) = purple and “neutral/non-salient” = gray. ZS and DC compared color coding and resolved most discrepancies through discussion and review of the original transcripts to help contextualize and ensure accuracy of summaries. In a few instances, discrepancies were resolved with input from additional study personnel.

Step 4: combine stakeholder valences for each factor (column code) within organizational units

ZS and DC applied rules for combining valences (see Fig. 10 of supplementary materials) from individual interviews within the same organizational unit to create a new row of summary valences for each organizational unit. For example, when one factor for a stakeholder had a “positive” valence and a different stakeholder had a “negative” valence, that factor became “mixed” for the organizational unit.

Step 5: copy factor valences from each organizational unit and order units by the outcome (i.e., optimization scores that were determined during process mapping)

Once the initial data matrices were completed, we combined summary valences from data matrix heat maps of each organizational unit into a single data matrix to compare contextual factors more easily across organizational units. DC copied all coded factors for each organizational unit into their respective CFIR domain to create a single data matrix containing the summary valences for all organizational units. One organizational unit was not included due to missing data, leaving 18 organizational units for analysis. Optimization scores were added to the matrix and organizational units were ordered from highest to lowest based on their optimization score. We then divided organizations into three groups based on natural breaks in the actual organizational optimization scores which grouped as follows: (1) score of “5″ = ”fully optimized,” (2) score of "0″ = “no program,” and (3) the remainder with scores ranging between “1″ to “3.5″ = ”non-optimized.”

Step 6: collapse related or overlapping codes within factors for data reduction and simplification

DC, with input from the research team, collapsed related CFIR factors. For example, “evidence strength and quality” and “relative advantage” were combined because they were often discussed together in stakeholder responses and usually had the same valence for each organizational unit. Specifically, the evidence that stakeholders often cited was how UTS did a better job at identifying LS compared to other approaches. As another example, most factors related to an organizational unit’s inner setting such as “networks and communications”, “culture”, “implementation climate”, and “readiness for implementation” were combined to create an overall inner setting valence. However, “structural characteristics” were not hypothesized to impact the outcome directly and were therefore not combined as part of overall setting (color-coding rules for collapsing factors can be seen in Fig. 12 within the supplementary materials for a visual representation).

Step 7: compare data across organizational units and create a final data matrix heat map of selected factors as a preparatory step for future analysis

A finalized and updated data matrix heat map with optimization scores from process mapping was reviewed to conduct cross-case comparisons and select CFIR factors. The color-coded visualization of the matrix heat map was used by the study team to hypothesize which factors may distinguish between optimized organizational units, non-optimized organizational units, and those without a UTS program. Factor selection was the final preparatory step before conducting coincidence analysis to formally test whether the hypothesized factors make a difference for the outcome as described in the initial IMPULSS study design [13].

Results

Process mapping results and interpretations

Process mapping revealed variation in protocols implemented to identify LS patients, both within and across organizations. UTS protocols included ten main steps which were partially impacted by differences in how institutions are structured and involvement of geographically distinct pathology groups.

Most (67%) of the healthcare organizations we interviewed had a single protocol for identifying LS, but three of the nine were divided into multiple organizational units. These findings led to the creation of 19 finalized process maps, one for each organizational unit identified (see Figs. 3 and 4 for examples of initial and reconciled/finalized versions, respectively). This visual representation assisted with systematically comparing protocols and detecting process gaps and inefficiencies related to inconsistent execution of the tumor screening protocol and referrals, lack of process automation, and no evidence of data tracking or quality assurance. These gaps and inefficiencies were related to five analogous optimization components that were used to calculate the optimization score:

  • Routine or consistent execution in conducting either immunohistochemistry (IHC) or microsatellite instability (MSI) testing. Note: Both methods are appropriate as an initial tumor screen to identify patients with LS [17] (IHC or MSI consistency)

  • The use of reflex testing using BRAF V600E and/or MLH1 promoter hypermethylation testing on a subset of “screen positive” tumors (MLH1 absent) as recommended by the NCCN [17] (routine reflex)

  • Consistency in referrals of patients with positive screens for germline genetic testing (referral consistency)

  • Systematic results tracking (evidence of tracking)

  • Ongoing quality assurance measures (quality assurance)

Fig. 3
figure 3

Section of the initial process map for organizational unit 1A. Key describes lines and symbols that were color-coded to represent varying stakeholder perspectives and to signify corroboration across stakeholders within an organization. Symbols were also used to identify and visually discern inconsistencies in stakeholder-reported interview data. Ten total stakeholders were interviewed from organization 1. Stakeholders not pictured were located at other organizational units (1B, 1C, and 1D)

Fig. 4
figure 4

Section of the reconciled process map for organizational unit 1A. Completion of process mapping resulted in 19 reconciled process maps, one for each organizational unit

The first three components of the optimization score (IHC or MSI consistency, routine reflex, and referral consistency) correlate to the CFIR process factor of “executing.” The last two components (evidence of tracking and quality assurance measures) fall within the CFIR process factor of “reflecting and evaluating.”

Four organizational units scored a 5 out of 5, indicating full UTS program optimization. The remaining 11 organizational units with a UTS program were not optimized. Given that scores for non-optimized programs were all between 1 and 3.5, all organizational units would all have been categorized the same regardless of whether we used 0.5 in our coding system. A total of four organizational units had no program and were included in the study to understand reasons why healthcare systems may not have a UTS program. Organizational units had varied optimization scores, including sites under a single organizational umbrella. As illustrated in Fig. 5, organization 4 had six units ranging from a score of 0 to 5. Likewise, we saw varied UTS protocol execution at organization 6 with scores ranging from 1 to 5.

Fig. 5
figure 5

Matrix of UTS protocol optimization levels by organizational unit. Presence of process optimization component = “1;” historical presence without current presence of optimization component, or unresolved discrepancy between stakeholders = “0.5;” absence of process optimization component = “0.” Optimization components were not applicable (N/a) at organizational units without a program and those received an overall optimization score of “0”

Data matrix heat mapping results and interpretations

The final heat map shown in Fig. 6 (also seen in Fig. 13 of supplemental materials) shows organizational units ordered based on their outcome scores from process mapping.

Fig. 6
figure 6

Final heat map showing the outcome and factors selected for coincidence analysis (CNA). Factor 1: evidence and relative advantage; Factor 2: cost; Factor 3: knowledge and attitudes of stakeholders; Factor 4: implementation champion; Factor 5: maintenance champion; Factor 6: planning and engaging stakeholders; Factor 7: inner setting (except structural); Factor 8: external networks (cosmopolitanism), peer pressure

Each row represents a single organizational unit with the selected color-coded factors organized in columns. Factors were selected by the research team because they were hypothesized to influence UTS implementation and/or optimization based on experience and literature review [11, 12, 15, 22, 31, 32]. The final data matrix heat map shows more blue patterns (presence of factor/facilitator/positive influence) seen at fully optimized sites, and more purple (mixed perceptions) and red (absence of a factor/presence of a barrier/negative influence) patterns at non-optimized sites and sites without a program.

The positive blue coloring in evidence and relative advantage (Factor 1) (Fig. 6) shows that 13 of 14 sites with a program, regardless of optimization level, had stakeholders who saw clear benefits and strong evidence favoring UTS; in contrast, purple coloring shows mixed perceptions among stakeholders at three of the four organizational units with no program. Cost (Factor 2) was included in our final factor selection because it was identified as the primary concern at one organizational unit and other literature discusses its importance to implementation [11, 31, 32]. Knowledge and attitudes of stakeholders (Factor 3) was included given that it showed a greater presence of blue among units with higher optimization scores.

The following quote illustrates the potential impacts of both knowledge and attitudes of stakeholders (Factor 3) and inner setting (Factor 7), “I think [UTS] is the kind of thing that [provider stakeholders] would be interested in [positive attitudes], so it may just be a matter of distribution of resources or where it’s going to fall with other priorities [inner setting]” — Genetic counselor, Unit 1D. Four of the five fully optimized sites had a positive inner setting. In contrast, programs that were not fully optimized had a mixed inner setting. Implementation champion (Factor 4) was selected due to prior research supporting the key role implementation champions can play [33,34,35] as well as our observation that this came up multiple times in qualitative data as evidenced by the following quote: “Make sure you have the right people at the table at the right time. I think that’s key, making sure you have the right [team] and identifying a champion, somebody who is really going to champion what you’re doing and show people why [UTS] is important” — Tumor registrar, Unit 3.

Evidence of a maintenance champion (Factor 5) was present at all the optimized sites. Even though a maintenance champion was also present at some non-optimized sites, qualitative data support the importance of a maintenance champion who is often not the same person as the implementation champion: “One of our other genetic counselors is taking some ownership of [UTS] now, but she has more regular touch-ins with those pathologists. A specific representative from the department is often helpful” — Genetic counselor, Unit 8.

Planning and engaging stakeholders (Factor 6) is logically necessary for implementation and optimization of UTS programs. In fact, only sites without a UTS program are lacking planning and engaging.

Lastly, cosmopolitanism (i.e., external networks) and peer pressure (Factor 8) was included as the final factor since stakeholders from multiple organizational units described their importance: “We address our algorithm every other year to reach out to [a leading healthcare organization in UTS] and the researchers to see where things are at and do review of the literature to update and see if we’re in line with [peers at leading organizations]” — Nurse practitioner, Unit 4A; “We emailed friends at other institutions. We weren’t starting from scratch, and you don’t have to. Don’t reinvent the wheel, just figure out what’s going on out there and what can work best for your institution” — Genetic counselor, Unit 5.

Discussion

In this methods manuscript, we described how two data visualization methods were useful in identifying and comparing factors that may be associated with the implementation and optimization of complex programs within healthcare organizations, specifically the implementation of UTS for LS. The first data visualization method, process mapping, used color-coded symbols to represent stakeholder descriptions of current UTS processes within their organization. Ecomapping [5] is a similar method that incorporates color-coded symbols to represent supportive care networks and individual stakeholder relationships, but process mapping helped to visualize and easily compare organizational processes. In our study, process mapping led to the identification and classification of distinct organizational units that we used to conduct cross-case comparisons, uncover similarities and differences across organizational units, and gain greater insight into organizational complexity impacting implementation variability.

Process mapping has been used previously for data visualization of UTS programs across multiple healthcare systems but was not used to document and reconcile stakeholder inconsistencies. Palter et al. [25] found one of the same components for the implementation optimization of UTS programs, namely ensuring the coordination of patient tracking to optimize the pathway for possible LS patients to meet with genetic counselors. The other key facilitators to successful implementation identified by Palter et al. were provider-stakeholder engagement and flexibility to tailor programs to relevant clinical sites. Our study expands on the findings of Palter et al. by identifying additional components for program optimization that we used to define implementation success and develop a scale measuring the degree to which a program has been optimally implemented and maintained.

Although the use of data matrices to make comparisons across sites is not new [2, 10] our study expanded its use by incorporating color-coded valences for factors believed to be important to implementation and optimization. We propose that this novel method of data matrix heat mapping offers a systematic approach to organize complex institutional data, assign valences, and conduct cross-site comparisons. One of the benefits of matrix heat mapping is that it allows researchers to see patterns within the final data matrix. Another benefit that matrix heat mapping shares with the multiple matrixed case study approach is that these approaches allow researchers to go back to the earlier data matrices to review additional details/summaries and easily gain a more nuanced or detailed understanding of each organizational unit. The earlier matrices can also be linked directly to supporting quotes. Furthermore, matrix heat mapping helped formalize a data consolidation technique for combining data from multiple stakeholders. Finally, data matrix heat mapping also provided a systematic approach to factor reduction, which is a necessary step before conducting other, more formal analyses such as coincidence analysis (CNA).

We plan to conduct CNA as the final analytic step in the IMPULSS study [13] to test whether and how the factors we selected using matrix heat mapping consistently make a difference for implementation and optimization. One advantage of CNA is its ability to identify whether the data support underlying causal complexity (e.g., more than one factor may be minimally necessary and sufficient for the outcome or multiple factors may form a causal chain that led to the outcome) [27, 36]. Currently, CNA is limited in the number of factors that can be successfully analyzed by the computer algorithm which builds models that fit the data and meet certain consistency and coverage thresholds. Others have used a data-driven methodology to select factors for CNA [37], and our matrix heat mapping approach provides another method to review and select factors for use in CNA.

Limitations

This study and the data visualization approaches described here have several limitations. First, data collection was limited to current stakeholders at the organizations in the IMPULSS study, some of whom were newer to the organization and lacked historical information related to the initial UTS program implementation. This may limit the ability to fully assess whether and how implementation champions impacted UTS program implementation at those organizational units. Several prior studies [33,34,35] have identified implementation champions as one of the necessary components for success, and several participants in our study described the important role champions played in their program implementation. However, we had one organizational unit that clearly lacked an implementation champion but had a non-optimized program due to their partnership with an external hospital that initiated tumor screening.

An additional limitation is that participant roles in the process of identifying LS and the number of stakeholders across organizational units varied. The study team attempted to bridge gaps in knowledge that were not obtained in the initial interviews by conducting member-checking with participants and their colleagues, but there were some organizational units where member-checking was not logistically possible.

Furthermore, although attempts were made to identify multiple stakeholders, a few organizational units had only one stakeholder perspective represented, especially among organizations that we ultimately divided into multiple units. Consequently, one site with a non-optimized program was excluded from the data matrix heat mapping analysis because the sole stakeholder interview provided limited information. The variability in stakeholders interviewed may have impacted results because certain CFIR factors were identified to be more likely assigned a valence of “non-salient” (color-coded gray) at organizational units where only a single stakeholder was interviewed. Factors deemed “non-salient” should not make a difference for the outcome because we expect that relevant factors would be mentioned in stakeholder interviews upon probing. Nevertheless, in some instances, these “non-salient” factors may represent missing perspectives of other stakeholders. The presence of multiple gray “non-salient” factors served as a secondary impetus for factor consolidation, although the primary impetus of consolidation was to conduct the planned CNA. The use of our consolidation process may obscure some of the nuanced details. For example, by combining aspects of the inner setting, it is not clear which ones are most critical. However, one benefit of our approach is that it allowed us to review earlier data matrices and we discovered that the most salient inner setting factors differed. For several organizational units, it was the presence of strong networks and communications, but for others, it was primarily the presence of available resources.

Additionally, program optimization was defined, in part, based on current NCCN guidelines [17], and guidelines may change as evidence evolves and healthcare organizations choose other strategies for LS identification such as tumor sequencing [38, 39] or direct-to-germline [39, 40] sequencing. Future studies following the methodological steps we describe will need to review current literature to assess for updates in guidelines since the implementation of new procedures could alter how optimization is defined or which factors are related to implementation and optimization.

While we did not directly test this approach against other methods, we identified limitations in existing methods. Specifically, while working to identify process optimization components, data complexity necessitated a way to consolidate and visualize data to readily identify, document, and resolve discrepancies across multiple stakeholder perspectives. Furthermore, the process maps made member checking much easier because it provided a quick way to visualize all key steps or gaps in each organization’s UTS process. In addition, we found that using the multiple matrixed method described by Kim et al. [2] did not meet our needs to consolidate data and highlight patterns to ultimately select factors for further analysis. This was the impetus for developing the color coding and consolidation process steps of matrix heat mapping.

This is the first study to apply these two data visualization techniques in tandem. Therefore, future studies that incorporate process mapping and data matrix heat mapping are needed to sufficiently validate this approach. Nevertheless, this work demonstrated proof of concept for using this approach in implementation science research.

Conclusion

We offer a detailed description of how two data visualization techniques were used to successfully organize and analyze data from an implementation study of a complex program (UTS) with multiple stakeholders and organizational units. Process mapping provided an efficient method to visualize protocols and operationalize the key optimization factors to create scores representing our implementation outcome. Matrix heat mapping using a priori CFIR factors helped identify local contextual factors impacting program implementation and optimization. Combining both the primary outcome (i.e., optimization scores from process mapping) and the summarized qualitative interview data formed the basis for our final “data matrix heat map.” These data visualization methods also helped consolidate factors that can be used for formal analysis using CNA to determine which of the factors differentiate between organizational units with an optimized program, a non-optimized program, or no program. These methodological approaches (or variations of them) may be useful for other implementation science studies aimed at understanding the impact of multiple organizational factors on implementation processes and outcomes.

Availability of data and materials

All data generated or analyzed during this study are included in this published article and its supplementary information files.

Abbreviations

NCI:

National Cancer Institute

LS:

Lynch syndrome

UTS:

Universal tumor screening

dMMR:

Mismatch repair deficiency

IHC:

Immunohistochemistry

MSI:

Microsatellite instability

NCCN:

National Comprehensive Cancer Network

ACOG:

American College of Obstetrics and Gynecology

ASCP:

American Society for Clinical Pathology

EGAPP:

Evaluation of Genomic Applications in Practice and Prevention

SGO:

Society of Gynecologic Oncology

QI:

Quality improvement

QA:

Quality assurance

CFIR:

Consolidated Framework for Implementation Research

CNA:

Coincidence analysis

References

  1. Perry CK, Damschroder LJ, Hemler JR, Woodson TT, Ono SS, Cohen DJ. Specifying and comparing implementation strategies across seven large implementation interventions: a practical application of theory. Implement Sci. 2019;14(1):32.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Kim B, Sullivan JL, Ritchie MJ, Connolly SL, Drummond KL, Miller CJ, et al. Comparing variations in implementation processes and influences across multiple sites: What works, for whom, and how? Psychiatry Res. 2020;1(283):112520.

    Article  Google Scholar 

  3. BarcinaLacosta T, Vulto AG, Turcu-Stiolica A, Huys I, Simoens S. Qualitative Analysis of the Design and Implementation of Benefit-Sharing Programs for Biologics Across Europe. BioDrugs. 2022;36(2):217–29.

    Article  Google Scholar 

  4. Silva HM, Gonzaga do Nascimento MM, de Morais Neves C, Oliveira IV, Cipolla CM, Batista de Oliveira GC, et al. Service blueprint of comprehensive medication management: A mapping for outpatient clinics. Res Social Adm Pharm. 2021;17(10):1727–36.

    Article  PubMed  Google Scholar 

  5. Ray RA, Street AF. Ecomapping: an innovative research tool for nurses. J Adv Nurs. 2005;50(5):545–52.

    Article  PubMed  Google Scholar 

  6. Ly S, Runacres F, Poon P. Journey mapping as a novel approach to healthcare: a qualitative mixed methods study in palliative care. BMC Health Serv Res. 2021;21(1):915.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Process mapping in healthcare: a systematic review | BMC Health Services Research | Full Text [Internet]. Available from: https://bmchealthservres.biomedcentral.com/articles/10.1186/s12913-021-06254-1. Cited 15 Jun 2022.

  8. Barter RL, Yu B. Superheat: An R package for creating beautiful and extendable heatmaps for visualizing complex data. J Comput Graph Stat. 2018;27(4):910–22.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Eisen MB, Spellman PT, Brown PO, Botstein D. Cluster analysis and display of genome-wide expression patterns. Proc Natl Acad Sci U S A. 1998;95(25):14863–8.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Gale NK, Heath G, Cameron E, Rashid S, Redwood S. Using the framework method for the analysis of qualitative data in multi-disciplinary health research. BMC Med Res Methodol. 2013;18(13):117.

    Article  Google Scholar 

  11. Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76.

    Article  PubMed  Google Scholar 

  12. Keith RE, Crosson JC, O’Malley AS, Cromp D, Taylor EF. Using the Consolidated Framework for Implementation Research (CFIR) to produce actionable findings: a rapid-cycle evaluation approach to improving implementation. Implement Sci. 2017;12(1):15.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Rahm AK, Cragun D, Hunter JE, Epstein MM, Lowery J, Lu CY, et al. Implementing universal Lynch syndrome screening (IMPULSS): protocol for a multi-site study to identify strategies to implement, adapt, and sustain genomic medicine programs in different organizational contexts. BMC Health Serv Res. 2018;30(18):824.

    Article  Google Scholar 

  14. Bhattacharya P, McHugh TW. Lynch Syndrome. In: StatPearls [Internet]. Treasure Island (FL): StatPearls Publishing; 2022 Available from: http://www.ncbi.nlm.nih.gov/books/NBK431096/. Cited 26 Aug 2022.

  15. Crain PR, Zepp JM, Gille S, Jenkins L, Kauffman TL, Shuster E, et al. Identifying patients with Lynch syndrome using a universal tumor screening program in an integrated healthcare system. Hered Cancer Clin Pract. 2022;20(1):17.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  16. Hunter JE, Zepp JM, Gilmore MJ, Davis JV, Esterberg EJ, Muessig KR, et al. Universal tumor screening for Lynch syndrome: Assessment of the perspectives of patients with colorectal cancer regarding benefits and barriers. Cancer. 2015;121(18):3281–9.

    Article  PubMed  Google Scholar 

  17. Genetic/Familial High-Risk Assessment: Colorectal | National Comprehensive Cancer Network | Full Text [Internet]. Available from: http://www.nccn.org/professionals/physician_gls/pdf/genetics_colon.pdfCited 15 Jun 2022.

  18. Hampel H. NCCN increases the emphasis on genetic/familial high-risk assessment in colorectal cancer. J Natl Compr Canc Netw. 2014;12(5 Suppl):829–31.

    Article  PubMed  Google Scholar 

  19. ACOG Practice Bulletin No. 147: Lynch syndrome. Obstet Gynecol. 2014;124(5):1042–54.

    Article  Google Scholar 

  20. Syngal S, Brand RE, Church JM, Giardiello FM, Hampel HL, Burt RW, et al. ACG clinical guideline: Genetic testing and management of hereditary gastrointestinal cancer syndromes. Am J Gastroenterol. 2015;110(2):223–62; quiz 263.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Sepulveda AR, Hamilton SR, Allegra CJ, Grody W, Cushman-Vokoun AM, Funkhouser WK, et al. Molecular Biomarkers for the Evaluation of Colorectal Cancer. Am J Clin Pathol. 2017;147(3):221–60.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  22. Cragun D, DeBate RD, Vadaparampil ST, Baldwin J, Hampel H, Pal T. Comparing universal Lynch syndrome tumor-screening programs to evaluate associations between implementation strategies and patient follow-through. Genet Med. 2014;16(10):773–82.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Hampel H, Pearlman R, Cragun D. Universal Tumor Screening for Lynch Syndrome. In: Valle L, Gruber SB, Capellá G, editors. Hereditary Colorectal Cancer: Genetic Basis and Clinical Implications [Internet]. Cham: Springer International Publishing; 2018 p. 233–55. Available from: https://doi.org/10.1007/978-3-319-74259-5_17. Cited 6 Jul 2022.

  24. The Consolidated Framework for Implementation Research – Technical Assistance for users of the CFIR framework [Internet]. [cited 2022 Aug 26]. Available from: https://cfirguide.org/.

  25. Palter VN, Baker NA, Pollett A, Daly C, Facey M, Rotenberg C, et al. Learning by Example: An International Perspective on Reflex-Testing for Lynch Syndrome. Ann Surg Oncol. 2019;26(2):425–36.

    Article  PubMed  Google Scholar 

  26. Baumgartner M, Ambühl M. Causal modeling with multi-value and fuzzy-set Coincidence Analysis. Polit Sci Res Methods. 2020;8(3):526–42.

    Article  Google Scholar 

  27. Wasmuth S, Belkiewitz J, Bravata D, Horsford C, Harris A, Smith C, et al. Protocol for evaluating external facilitation as a strategy to nationally implement a novel stigma reduction training tool for healthcare providers. Implement Sci Commun. 2022;12(3):88.

    Article  Google Scholar 

  28. Lincoln YS, Guba EG, Guba 19..-2008 EG. Naturalistic Inquiry. SAGE; 1985. 422 .

  29. Welcome to RQDA Project [Internet]. Available from: https://rqda.r-forge.r-project.org/. Cited 26 Aug 2022.

  30. Damschroder LJ, Reardon CM, Sperber N, Robinson CH, Fickel JJ, Oddone EZ. Implementation evaluation of the Telephone Lifestyle Coaching (TLC) program: organizational factors associated with successful implementation. Transl Behav Med. 2017;7(2):233–41.

    Article  PubMed  Google Scholar 

  31. Roberts SLE, Healey A, Sevdalis N. Use of health economic evaluation in the implementation and improvement science fields-a systematic literature review. Implement Sci. 2019;14(1):72.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Gold HT, McDermott C, Hoomans T, Wagner TH. Cost data in implementation science: categories and approaches to costing. Implement Sci. 2022;17(1):11.

    Article  PubMed  PubMed Central  Google Scholar 

  33. Inside help: An integrative review of champions in healthcare-related implementation - Edward J Miech, Nicholas A Rattray, Mindy E Flanagan, Laura Damschroder, Arlene A Schmid, Teresa M Damush. 2018 [Internet]. [cited 2022 Jul 6]. Available from: https://journals.sagepub.com/doi/full/10.1177/2050312118773261.

  34. Shortell SM, Marsteller JA, Lin M, Pearson ML, Wu SY, Mendel P, et al. The role of perceived team effectiveness in improving chronic illness care. Med Care. 2004;42(11):1040–8.

    Article  PubMed  Google Scholar 

  35. Bonawitz K, Wetmore M, Heisler M, Dalton VK, Damschroder LJ, Forman J, et al. Champions in context: which attributes matter for change efforts in healthcare? Implementation Science. 2020;15(1):62.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Whitaker RG, Sperber N, Baumgartner M, Thiem A, Cragun D, Damschroder L, et al. Coincidence analysis: a new method for causal inference in implementation science. Implementation Science. 2020;15(1):108.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Strategy Configurations Directly Linked to Higher Hepatitis... : Medical Care [Internet]. [cited 2022 Aug 30]. Available from: https://journals.lww.com/lww-medicalcare/fulltext/2020/05000/strategy_configurations_directly_linked_to_higher.12.aspx.

  38. Hampel H, Pearlman R, Beightol M, Zhao W, Jones D, Frankel WL, et al. Assessment of Tumor Sequencing as a Replacement for Lynch Syndrome Screening and Current Molecular Tests for Patients With Colorectal Cancer. JAMA Oncol. 2018;4(6):806–13.

    Article  PubMed  Google Scholar 

  39. Hao J, Hassen D, Gudgeon JM, Snyder SR, Hampel H, Williams MS, et al. Economic Evaluation of Universal Lynch Syndrome Screening Protocols among Newly Diagnosed Patients with Colorectal Cancer. J Pers Med. 2021;11(12):1284.

    Article  PubMed  PubMed Central  Google Scholar 

  40. Direct-to-sequencing vs immunohistochemistry protocol for Lynch Syndrome? PharmacoEcon Outcomes News. 2021;878(1):9–9.

Download references

Acknowledgements

The collaborations of healthcare systems participating in this study are possible through the Healthcare System Research Network (HCSRN) and the Cancer Research Network (CRN). Geisinger, KPNW, KPCO, MPCI, PAMFRI, Harvard Pilgrim Health Care, and HealthPartners are all members of the HCSRN and the CRN. CHI is a previous member of the HCSRN. RS is a CRN Scholar. Additional guidance on study activities and dissemination is provided in collaboration with the Lynch Syndrome Screening Network (LSSN).

Funding

This project and publication are supported by the National Cancer Institute (NCI) twenty-first Century Cures Act—Beau Biden Cancer Moonshot R01CA211723. The content of this article is the sole responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Author information

Authors and Affiliations

Authors

Contributions

Interviews were conducted by ZS, IL, VS, and JA; ZS, DC, and AKR analyzed and interpreted process data; ZS created process maps for each organizational unit; DC and JBA analyzed and interpreted data using the CFIR; AKR, JLS, MME, CYL, PAP, RNS, SYL, ANBH, and JEH provided site-specific data, assisted with stakeholder recruitment, and arranged site member-checking; DC created heat maps for the data matrix heat map approach and ZS corroborated findings. MSW assisted with the initial study design and provided direction for recruitment and data collection methods. ZS prepared the manuscript, and DC, AKR, and JLS were major contributors in writing the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Zachary M. Salvati.

Ethics declarations

Ethics approval and consent to participate

Members of Geisinger’s Institutional Review Board (IRB) reviewed the research protocol under [45 CFR 46.110(b)(1) and/or 21 CFR 56.110(b)(1)] expedited review Category 7: Research on individual or group characteristics or behavior (including, but not limited to, research on perception, cognition, motivation, identity, language, communication, cultural beliefs or practices, and social behavior) or research employing survey, interview, oral history, focus group, program evaluation, human factors evaluation, or quality assurance methodologies and approved it with minor modifications on 05/21/2017, which were administratively approved on 08/08/2017.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1:

Fig. 1. Data Extraction Guide for Universal Tumor Screening (UTS) Protocol Processes. Fig 2. Process and Contextual Differences within Organization 1. Fig 3. Initial Process Map for Organizational Unit 1A. Fig 4. Reconciled Process Map for Organizational Unit 1A. Fig. 5. Process Gap/Inefficiency to Optimization Component Conversion Table. Fig. 6. Matrix of UTS Protocol Optimization Levels by Organizational Unit. Fig. 7. Section of the CFIR Codebook Used for Qualitative Data Analysis. Fig. 8. Data Matrix of Factors Related to Intervention Characteristics from Organizational Unit 7. Fig. 9. Example of a Data Matrix Heat Map for Factors Related to Intervention Characteristics. Fig. 10. Example of Combined Codes for Factors Related to Intervention Characteristics. Fig. 11. Consolidated Data Matrix Heat Map for Characteristics of Intervention by Optimization Score. Fig. 12. Example of Collapsed Code ‘Evidence & Relative Advantage’ for all Organizational Units. Fig. 13. Final Heat Map of Factors Selected for Coincidence Analysis (CNA).

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Salvati, Z.M., Rahm, A.K., Williams, M.S. et al. A picture is worth a thousand words: advancing the use of visualization tools in implementation science through process mapping and matrix heat mapping. Implement Sci Commun 4, 43 (2023). https://doi.org/10.1186/s43058-023-00424-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s43058-023-00424-4

Keywords