Measuring Agreement with Cohen's Kappa Statistic | by Blake Samaha | Towards Data Science
Simple guide to confusion matrix terminology
Metrics to evaluate classification models with R codes: Confusion Matrix, Sensitivity, Specificity, Cohen's Kappa Value, Mcnemar's Test - Data Science Vidhya
Explaining the unsuitability of the kappa coefficient in the assessment and comparison of the accuracy of thematic maps obtained by image classification - ScienceDirect
Decoding the Confusion Matrix - KeyToDataScience
The accuracy and Cohen's kappa of the confusion matrix example for the... | Download Table