Skip to main content

Table 5 Categorical concordance

From: Quantifying implementation strategy and dissemination channel preferences and experiences for pain management in primary care: a novel implementer-reported outcome

Cohen’s kappa (κ)

Categorical interpretation

Dissemination concordance

n (%) [M]

Implementation concordance

n (%) [M]

−1

Perfect disagreement

  

−0.81 to −0.99

Near perfect disagreement

  

−0.61 to −0.80

Substantial disagreement

  

−0.41to −0.60

Moderate disagreement

2 (2.0) [−0.47]

 

−0.21 to −0.40

Fair disagreement

3 (3.0) [−0.29]

6 (5.9) [−0.30]

−0.10 to −0.20

Slight disagreement

2 (2.0) [−0.15]

8 (7.9) [−0.15]

−0.09–0.09

No different from chance

5 (4.9) [0.01]

24 (23.8) [0.03]

0.10–0.20

Slight agreement

10 (9.9) [0.13]

23 (22.8) [0.15]

0.21–0.40

Fair agreement

20 (19.8) [0.31]

25 (24.7) [0.29]

0.41–0.60

Moderate agreement

24 (23.8) [0.51]

9 (8.9) [0.50]

0.61–0.80

Substantial agreement

19 (18.8) [0.70]

3 (3.0) [0.65]

0.81–0.99

Near perfect agreement

14 (13.9) [0.88]

1 (1.0) [0.85]

1

Perfect agreement

2 (2.0) [1.0]

2 (2.0) [1.0]

  1. Table 5 shows the seven-level categorization of Cohen’s kappa (κ) at the individual level. Dissemination and implementation concordance were calculated for each participant and categorized. We reported the number of participants, the percent of total, and average for each category