Skip to main content

Table 3 Inter-rater agreement (percentage agreement) and reliability (kappa coefficients) of the items from the COSMIN checklist (COSMIN step 4)

From: Inter-rater agreement and reliability of the COSMIN (COnsensus-based Standards for the selection of health status Measurement Instruments) Checklist

Item nr

Item

N (minus articles with 1 rating)a

% agreement

N

Kappa

Generalisability Box (n = 866) b

c

   

Was the sample in which the HR-PRO instruments was evaluated adequately described? In terms of:

    

1d

median or mean age (with standard deviation or range)?

733

86

865

0.36

2d

distribution of sex?

735

88

863

0.38e

3

important disease characteristics (e.g. severity, status, duration) and description of treatment?

746

80

862

0.39f

4d

setting(s) in which the study was conducted? e.g. general population, primary care or hospital/rehabilitation care

735

89

863

0.30e

5d

countries in which the study was conducted?

733

90

861

0.40e

6d

language in which the HR-PRO instrument was evaluated?

733

86

861

0.41e

7d

Was the method used to select patients adequately described? e.g. convenience, consecutive, or random

729

81

857

0.40

8

Was the percentage of missing responses (response rate) acceptable?

724

82

849

0.48

  1. a When calculating percentage agreement, articles that were only scored once on the particular item were not taken into account;b number of times a box was evaluated;c sample sizes of Generalisability box are much higher that other items, because scores of the items on the Generalisability box for all measurement properties were combined;d dichotomous item;e Items with low dispersal i.e. more than 75% of the raters who responded to an item rated the same response category;f Combined kappa coefficient calculated because of nominal response scale in a one-way design; printed in bold indicates Kappa > 0.70 or % agreement >80%.