====================================================================== INTER-ANNOTATOR AGREEMENT SUMMARY REPORT ====================================================================== Dataset Information: ---------------------------------------------------------------------- Total Samples Evaluated: 2,000 triplets Total Annotation Rows: 8,000 (2,000 × 4 annotators) Annotators: 4 (BH1, BH2, HT1, HT2) Rating Scale: 1-5 (Likert scale) Evaluation Criteria: 4 dimensions ====================================================================== FLEISS' KAPPA (Inter-Annotator Agreement) ====================================================================== Translation Accuracy κ = 0.816 (Almost Perfect Agreement) Semantic Equivalence κ = 0.812 (Almost Perfect Agreement) Grammatical Correctness κ = 0.836 (Almost Perfect Agreement) Literary Tone Preservation κ = 0.853 (Almost Perfect Agreement) Overall (Mean) κ = 0.829 (Almost Perfect Agreement) ====================================================================== INTRACLASS CORRELATION COEFFICIENT (ICC) ====================================================================== Translation Accuracy ICC = 0.949 Semantic Equivalence ICC = 0.948 Grammatical Correctness ICC = 0.955 Literary Tone Preservation ICC = 0.957 Overall (Mean) ICC = 0.952 ====================================================================== COMPOSITE FLUENCY STATISTICS ====================================================================== Mean Composite Fluency: 4.13 ± 0.51 Range: 2.48 to 5.00 Average Std Deviation: 0.05 Quality Distribution: Excellent (≥4.5): 676 (33.8%) Good (4.0-4.5): 674 (33.7%) Acceptable (3.5-4.0): 458 (22.9%) Borderline (<3.5): 192 (9.6%) ====================================================================== INTERPRETATION (Landis & Koch, 1977) ====================================================================== Kappa Interpretation Scale: < 0.00: Poor agreement 0.00-0.20: Slight agreement 0.21-0.40: Fair agreement 0.41-0.60: Moderate agreement 0.61-0.80: Substantial agreement 0.81-1.00: Almost perfect agreement ← BHT25 (κ = 0.829) The corpus-wide weighted composite fluency was 4.13 ± 0.51." ====================================================================== KAPPA CALCULATION EXPLANATION ====================================================================== Fleiss' Kappa measures the degree of agreement among multiple raters when assigning categorical ratings to a number of items. Formula: κ = (P̄ - Pₑ) / (1 - Pₑ) Where: P̄ = Observed proportion of agreement among raters Pₑ = Expected proportion of agreement by chance For our dataset: - 2,000 triplets rated by 4 annotators each - 5-point Likert scale (ratings 1-5) - κ = 0.83 means annotators agreed 83% more often than expected by chance Example: If all 4 annotators give rating '5': Perfect agreement (P = 1.0) If ratings are 5, 5, 5, 4: High agreement (P ≈ 0.75) If ratings are 5, 4, 3, 2: Low agreement (P ≈ 0.0) The achieved κ = 0.829 falls in the 'Almost Perfect Agreement' category.