“…Finally, in 1971, Kappa was modified by Fleiss to allow the reproducibility's measurement in cases where several (more than two) observers are judging cases. During the last decades, several papers have appeared in the medical and psychological literature, which discuss Kappa indexes for measuring agreement between two or more raters (Brilliant, Lepowski, & Musch, 1983;Kraemer, 1992;Krummenauer, 2000;Little, Worthingham-Roberts, & Mann, 1984;Posner, Sampson, Caplan, Ward, & Cheney, 1990;Robert & McNaemee, 1998). However, the Fleiss index for multiple raters is unweighted (as the original Kappa presented by Cohen [1960]); hence, it treats all disagreements equally.…”