File size: 3,371 Bytes
46b01b4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
======================================================================
INTER-ANNOTATOR AGREEMENT SUMMARY REPORT
======================================================================

Dataset Information:
----------------------------------------------------------------------
Total Samples Evaluated: 2,000 triplets
Total Annotation Rows: 8,000 (2,000 × 4 annotators)
Annotators: 4 (BH1, BH2, HT1, HT2)
Rating Scale: 1-5 (Likert scale)
Evaluation Criteria: 4 dimensions


======================================================================
FLEISS' KAPPA (Inter-Annotator Agreement)
======================================================================

Translation Accuracy                κ = 0.816  (Almost Perfect Agreement)
Semantic Equivalence                κ = 0.812  (Almost Perfect Agreement)
Grammatical Correctness             κ = 0.836  (Almost Perfect Agreement)
Literary Tone Preservation          κ = 0.853  (Almost Perfect Agreement)
Overall (Mean)                      κ = 0.829  (Almost Perfect Agreement)

======================================================================
INTRACLASS CORRELATION COEFFICIENT (ICC)
======================================================================

Translation Accuracy                ICC = 0.949
Semantic Equivalence                ICC = 0.948
Grammatical Correctness             ICC = 0.955
Literary Tone Preservation          ICC = 0.957
Overall (Mean)                      ICC = 0.952

======================================================================
COMPOSITE FLUENCY STATISTICS
======================================================================

Mean Composite Fluency:           4.13 ± 0.51
Range:                            2.48 to 5.00
Average Std Deviation:            0.05

Quality Distribution:
  Excellent (≥4.5):               676 (33.8%)
  Good (4.0-4.5):                 674 (33.7%)
  Acceptable (3.5-4.0):           458 (22.9%)
  Borderline (<3.5):              192 (9.6%)

======================================================================
INTERPRETATION (Landis & Koch, 1977)
======================================================================

Kappa Interpretation Scale:
  < 0.00:      Poor agreement
  0.00-0.20:   Slight agreement
  0.21-0.40:   Fair agreement
  0.41-0.60:   Moderate agreement
  0.61-0.80:   Substantial agreement
  0.81-1.00:   Almost perfect agreement  ← BHT25 (κ = 0.829)


The corpus-wide weighted composite fluency was 4.13 ± 0.51."

======================================================================
KAPPA CALCULATION EXPLANATION
======================================================================

Fleiss' Kappa measures the degree of agreement among multiple raters
when assigning categorical ratings to a number of items.

Formula: κ = (P̄ - Pₑ) / (1 - Pₑ)

Where:
  P̄ = Observed proportion of agreement among raters
  Pₑ = Expected proportion of agreement by chance

For our dataset:
  - 2,000 triplets rated by 4 annotators each
  - 5-point Likert scale (ratings 1-5)
  - κ = 0.83 means annotators agreed 83% more often than expected by chance

Example:
  If all 4 annotators give rating '5': Perfect agreement (P = 1.0)
  If ratings are 5, 5, 5, 4: High agreement (P ≈ 0.75)
  If ratings are 5, 4, 3, 2: Low agreement (P ≈ 0.0)

The achieved κ = 0.829 falls in the 'Almost Perfect Agreement' category.