kappa {lrd} | R Documentation |
Cohen's Kappa
Description
This function returns Cohen's Kappa k for two raters. Kappa indicates the inter-rater reliability for categorical items. High scores (closer to one) indicate agreement between raters, while low scores (closer to zero) indicate low agreement between raters. Negative numbers indicate they don't agree at all!
Usage
kappa(rater1, rater2, confidence = 0.95)
Arguments
rater1 |
Rater 1 scores or categorical listings |
rater2 |
Rater 2 scores or categorical listings |
confidence |
Confidence interval proportion for the kappa interval estimate. You must supply a value between 0 and 1. |
Details
Note: All missing values will be ignored. This function calculates kappa for 0 and 1 scoring. If you pass categorical variables, the function will return a percent match score between these values.
Value
p_agree |
Percent agreement between raters |
kappa |
Cohen's kappa for yes/no matching |
se_kappa |
Standard error for kappa wherein standard error is the square root of: (agree \* (1-agree)) / (N \* (1 - random agreement)^2) |
kappa_LL |
Lower limit for the confidence interval of kappa |
kappa_UL |
Upper limit for the confidence interval of kappa |
Examples
#This dataset includes two raters who wrote the word listed by
#the participant and rated if the word was correct in the recall
#experiment.
data(rater_data)
#Consider normalizing the text if raters used different styles
#Calculate percent match for categorical answers
kappa(rater_data$rater1_word, rater_data$rater2_word)
kappa(rater_data$rater1_score, rater_data$rater2_score)