Wajinimi commited on
Commit
d7f1a72
·
verified ·
1 Parent(s): f478a26

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +202 -4
README.md CHANGED
@@ -1,5 +1,203 @@
1
  ---
2
- license: ecl-2.0
3
- size_categories:
4
- - 10K<n<100K
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ license: cc-by-4.0
5
+ pretty_name: MAMA Communicative Intent Dataset (INCA-A Annotated)
6
+ tags:
7
+ - NLP
8
+ - child-language
9
+ - intent-classification
10
+ - conversational-ai
11
+ - developmental-linguistics
12
+ - dialogue
13
+ - child-speech
14
+ task_categories:
15
+ - text-classification
16
+ task_ids:
17
+ - intent-classification
18
+ ---
19
+
20
+ # MAMA Communicative Intent Dataset (INCA-A Annotated)
21
+
22
+ ## Overview
23
+
24
+ The **MAMA Communicative Intent Dataset** is a linguistically annotated corpus of child utterances designed to support research in **child-centred Natural Language Processing (NLP)** and **communicative intent recognition in early language development**.
25
+
26
+ The dataset contains **10,800 child utterances** annotated using the **INCA Communicative Coding System** (Ninio et al., 1994), a developmental framework that identifies the communicative functions underlying children's speech.
27
+
28
+ To make the dataset suitable for machine learning, the original INCA codes were mapped to **23 refined intent categories** representing distinct communicative behaviours in early child language.
29
+
30
+ This dataset was created as part of the **MAMA (Machine-Assisted Maternal Assistant)** research project, which investigates how artificial intelligence systems can better understand the communicative behaviour of young children.
31
+
32
+ Unlike many NLP corpora that normalise or correct non-standard language, this dataset **preserves authentic developmental linguistic features**, including telegraphic speech and missing grammatical markers.
33
+
34
+ ---
35
+
36
+ # Dataset Summary
37
+
38
+ | Property | Value |
39
+ |--------|------|
40
+ | Total utterances | 10,800 |
41
+ | Total labelled instances | 11,410 |
42
+ | Intent categories | 23 |
43
+ | Annotation framework | INCA Communicative Coding System |
44
+ | Source corpus | CHILDES |
45
+ | Language | English |
46
+ | Task | Intent Classification |
47
+
48
+ The dataset reflects **naturalistic child language**, resulting in class imbalance typical of real-world conversational data.
49
+
50
+ Example distribution:
51
+
52
+ | Intent | Frequency |
53
+ |------|------|
54
+ | Observation / Reference | 2,964 |
55
+ | Narrative / Storytelling | 1,664 |
56
+ | Comfort | 4 |
57
+
58
+ ---
59
+
60
+ # Annotation Framework
61
+
62
+ The dataset is based on the **INCA Communicative Coding System**, introduced in:
63
+
64
+ > Ninio, A., Snow, C., Pan, B., & Rollins, P. (1994).
65
+ > *Classifying communicative acts in children's interactions.*
66
+
67
+ The INCA system categorises **communicative functions** in children's speech rather than grammatical structure alone.
68
+
69
+ In this dataset, INCA codes were mapped into **refined NLP intent categories** suitable for supervised machine learning.
70
+
71
+ ---
72
+
73
+ # Mapping from INCA Codes to Refined Intent Categories
74
+
75
+ | INCA Category | INCA Code | Refined Intent Category |
76
+ |---------------|----------|-------------------------|
77
+ | Directing hearer’s attention | DHA / CL | Attention |
78
+ | Speech elicitation | EI, RT, EA | Imitation |
79
+ | Questions | QN, YQ, TQ | Question |
80
+ | Evaluation | ET | Excitement |
81
+ | Discussing related-to-present | DRP | Narrative or Storytelling |
82
+ | Discussing joint focus | DJF | Observation or Reference |
83
+ | Statements | WS | Desire or Action |
84
+ | Negotiating activity | NIA / DW | Disagreement or Correction |
85
+ | Marking | MRK | Gratitude |
86
+ | Comforting | CMO | Comfort |
87
+ | Directiveness | RP | Request |
88
+ | Directiveness | RD, CS | Refusal |
89
+ | Directiveness | GR | Explanation or Justification |
90
+ | Declaration | YD / AP | Agreement or Acknowledgment |
91
+ | Marking | MK | Greeting |
92
+ | Marking | EM | Distress or Pain |
93
+ | Marking | EN | Emotion |
94
+ | Fantasy discussion | DFW | Playtalk or Fantasy |
95
+ | Possession negotiation | PSS | Possession |
96
+ | Request / Suggest | RP | Need |
97
+ | Dare / Challenge | DR | Command |
98
+ | Disapprove / Protest | DS, ED, DW | Complaint |
99
+
100
+ ---
101
+
102
+ # Annotation Protocol
103
+
104
+ Annotation followed a **two-stage validation procedure**.
105
+
106
+ ### Stage 1 — Initial Annotation
107
+
108
+ All utterances were initially labelled by the **primary researcher** using the INCA communicative coding framework.
109
+
110
+ ### Stage 2 — Expert Re-annotation
111
+
112
+ To strengthen validity, the dataset was independently reviewed by two domain experts:
113
+
114
+ - **Developmental Psychologist**
115
+ - **Experienced Early-Years Teacher**
116
+
117
+ This ensured both **developmental theoretical grounding** and **practical child-language expertise**.
118
+
119
+ ---
120
+
121
+ # Inter-Annotator Reliability
122
+
123
+ Agreement between annotators was measured using **Cohen's Kappa (κ)**.
124
+
125
+ ### Observed Agreement
126
+
127
+ \[
128
+ P_o = \frac{\sum C_{ii}}{N}
129
+ \]
130
+
131
+ Where:
132
+
133
+ - \(C_{ii}\) = number of rows where annotators assigned the same category
134
+ - \(N\) = total number of annotated rows
135
+
136
+ ### Cohen's Kappa
137
+
138
+ \[
139
+ \kappa = \frac{P_o - P_e}{1 - P_e}
140
+ \]
141
+
142
+ Where expected agreement is defined as:
143
+
144
+ \[
145
+ P_e = \sum \left(\frac{R_i}{N} \cdot \frac{C_i}{N}\right)
146
+ \]
147
+
148
+ Where:
149
+
150
+ - \(R_i\) = rows assigned to category \(i\) by annotator 1
151
+ - \(C_i\) = rows assigned to category \(i\) by annotator 2
152
+
153
+ The resulting score was:
154
+
155
+ **κ = 0.81**
156
+
157
+ According to **Landis and Koch (1977)**, this represents **almost perfect agreement**, indicating strong reliability in intent categorisation.
158
+
159
+ ---
160
+
161
+ # Linguistic Characteristics
162
+
163
+ ## Average Utterance Length
164
+
165
+ Average utterance length was computed as:
166
+
167
+ \[
168
+ \text{Average utterance length} =
169
+ \frac{\sum_{i=1}^{N} (\text{token length of utterance}_i)}{N}
170
+ \]
171
+
172
+ Analysis revealed that:
173
+
174
+ - **Explanation or Justification**
175
+ - **Narrative or Storytelling**
176
+ - **Desire or Action**
177
+
178
+ tend to produce **longer utterances**, indicating more verbose communicative behaviour.
179
+
180
+ In contrast:
181
+
182
+ - **Observation or Reference**
183
+
184
+ typically contains **shorter utterances**, reflecting concise descriptions of objects or events in the shared environment.
185
+
186
+ ---
187
+
188
+ ## Lexical Diversity
189
+
190
+ Lexical diversity analysis showed variation across communicative intents.
191
+
192
+ Categories such as:
193
+
194
+ - **Observation**
195
+ - **Narrative or Storytelling**
196
+
197
+ exhibited **higher vocabulary diversity**, reflecting descriptive language use.
198
+
199
+ Conversely:
200
+
201
+ - **Agreement or Acknowledgement**
202
+
203
+ showed **low lexical diversity**, as these responses often rely on short, formulaic expressions such as: