What is a good inter-rater reliability score?

A good inter-rater reliability score depends on the statistic used and the context of the study.

For Cohen’s kappa (two raters), common guidelines are:

  • < 0.20: Poor agreement
  • 0.21–0.40: Fair agreement
  • 0.41–0.60: Moderate agreement
  • 0.61–0.80: Substantial agreement
  • 0.81–1.00: Almost perfect agreement

For the Intraclass Correlation Coefficient (interval or ratio data), similar thresholds are used:

  • < 0.50: Poor agreement
  • 0.51–0.75: Moderate agreement
  • 0.76–0.90: Good agreement
  • > 0.91: Excellent agreement