Cohen's Kappa: What it is, when to use it, and how to avoid its pitfalls | by Rosaria Silipo | Towards Data Science
![Beyond kappa: an informational index for diagnostic agreement in dichotomous and multivalue ordered-categorical ratings | Medical & Biological Engineering & Computing Beyond kappa: an informational index for diagnostic agreement in dichotomous and multivalue ordered-categorical ratings | Medical & Biological Engineering & Computing](https://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11517-020-02261-2/MediaObjects/11517_2020_2261_Figd_HTML.png)
Beyond kappa: an informational index for diagnostic agreement in dichotomous and multivalue ordered-categorical ratings | Medical & Biological Engineering & Computing
![Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics](https://statistics.laerd.com/spss-tutorials/img/ck/crosstabs-cells-dialogue-box-cohens-kappa-v27.png)
Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics
![Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics](https://statistics.laerd.com/spss-tutorials/img/ck/cohens-kappa-symmetric-measures-table-v27.png)
Cohen's kappa in SPSS Statistics - Procedure, output and interpretation of the output using a relevant example | Laerd Statistics
![Symmetry | Free Full-Text | An Empirical Comparative Assessment of Inter-Rater Agreement of Binary Outcomes and Multiple Raters Symmetry | Free Full-Text | An Empirical Comparative Assessment of Inter-Rater Agreement of Binary Outcomes and Multiple Raters](https://www.mdpi.com/symmetry/symmetry-14-00262/article_deploy/html/images/symmetry-14-00262-g001.png)
Symmetry | Free Full-Text | An Empirical Comparative Assessment of Inter-Rater Agreement of Binary Outcomes and Multiple Raters
![AgreeStat/360: computing agreement coefficients for 2 raters (Cohen's kappa, Gwet's AC1/AC2, Krippendorff's alpha, and more) based on raw ratings AgreeStat/360: computing agreement coefficients for 2 raters (Cohen's kappa, Gwet's AC1/AC2, Krippendorff's alpha, and more) based on raw ratings](https://www.agreestat.com/examples/pictures/cac_data_raw2rr.png)