zentavious commited on
Commit
eb2ca26
·
verified ·
1 Parent(s): b1e5de8

Sensitive info and bias sections

Browse files
Files changed (1) hide show
  1. README.md +10 -18
README.md CHANGED
@@ -276,7 +276,7 @@ This dataset originates from conversations between Eedi tutors and students to s
276
 
277
  To facilitate **downsampling**, talk moves were applied using the GPT-based classification model in [Moreau-Pernet et al. (2024)](https://dl.acm.org/doi/10.1145/3657604.3664664). The model was fine-tuned on conversation transcripts from small-group math tutoring sessions.
278
 
279
- To evaluate the model’s generalizability to 1:1 chat-based math tutoring sessions, we manually annotated a validation set of 200 tutor utterances sampled through weighted stratified sampling using the orignal codebook from [Moreau-Pernet et al. (2024)](https://dl.acm.org/doi/10.1145/3657604.3664664). The sample distribution was flattened by 0.8 of the original label distribution represented in the full 4k dialogue dataset in order to validate more examples of minority class labels.
280
 
281
  An error analysis was conducted on all mismatched labels. **All** instances where the model predicted a `<Getting Student to Relate>` label involved a common type of Eedi's word problems involving a conversation between two fictional students. The original codebook was not designed with this unique edge-case in mind and as such we posit the `<Getting Student to Relate>` label may not generalize this to 1:1 context. Considering this limitation, we present an additional results excluding GSR labels and find similar acceptable metrics to the original paper.
282
 
@@ -320,38 +320,30 @@ We make three important distinctions:
320
 
321
  3. **NER Model release constraints**: Due to the sensitive nature of the labeling task, the fine-tuned NER model used in this process cannot be released publicly due to re-identification concerns. It could potentially be used to detect residual PII in datasets processed using the HIPS anonymization method.
322
 
323
- #### Who are the annotators?
324
-
325
- <!-- This section describes the people or systems who created the annotations. -->
326
-
327
- [More Information Needed]
328
-
329
  #### Personal and Sensitive Information
330
 
331
  <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
332
 
333
- [More Information Needed]
334
 
335
  ## Bias, Risks, and Limitations
336
 
337
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
338
 
339
- [More Information Needed]
340
-
341
- ### Recommendations
342
-
343
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
344
-
345
- Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
346
 
347
  ## Citation [optional]
348
 
349
- Publication in Progress. Please reach out to Matthew.Zent@eedi.com for citation information.
350
 
351
- ## Dataset Card Authors
352
 
353
  Matthew Zent
354
 
355
  ## Dataset Card Contact
356
 
357
- Matthew.Zent@eedi.com
 
276
 
277
  To facilitate **downsampling**, talk moves were applied using the GPT-based classification model in [Moreau-Pernet et al. (2024)](https://dl.acm.org/doi/10.1145/3657604.3664664). The model was fine-tuned on conversation transcripts from small-group math tutoring sessions.
278
 
279
+ To evaluate the model’s generalizability to 1:1 chat-based math tutoring sessions, the 1st author manually annotated a validation set of 200 tutor utterances sampled through weighted stratified sampling using the orignal codebook from [Moreau-Pernet et al. (2024)](https://dl.acm.org/doi/10.1145/3657604.3664664). The sample distribution was flattened by 0.8 of the original label distribution represented in the full 4k dialogue dataset in order to validate more examples of minority class labels.
280
 
281
  An error analysis was conducted on all mismatched labels. **All** instances where the model predicted a `<Getting Student to Relate>` label involved a common type of Eedi's word problems involving a conversation between two fictional students. The original codebook was not designed with this unique edge-case in mind and as such we posit the `<Getting Student to Relate>` label may not generalize this to 1:1 context. Considering this limitation, we present an additional results excluding GSR labels and find similar acceptable metrics to the original paper.
282
 
 
320
 
321
  3. **NER Model release constraints**: Due to the sensitive nature of the labeling task, the fine-tuned NER model used in this process cannot be released publicly due to re-identification concerns. It could potentially be used to detect residual PII in datasets processed using the HIPS anonymization method.
322
 
 
 
 
 
 
 
323
  #### Personal and Sensitive Information
324
 
325
  <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
326
 
327
+ This dataset contains real student-tutor dialogues. Extensive efforts were made to mitigate privacy risks, but these risks cannot be alleviated completely. Identifying details were removed or anonymized where detected, but some sensitive content may persist. If you have concerns about any content, please contact the first author at [matthew.zent@eedi.co.uk](mailto:matthew.zent@eedi.co.uk).
328
 
329
  ## Bias, Risks, and Limitations
330
 
331
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
332
 
333
+ We present aggregate demographic and behavioral statistics to help downstream users assess potential biases in the dataset. Here we note highlight high-impact biases and limitations for downstream use:
334
+ 1. QATD’s focuses on UK students and intentionally excludes US-based learners.
335
+ 2. The data reflects real interactions on the Eedi platform, where tutors sometimes manage multiple students simultaneously, especially during peak hours.
336
+ 3. The collection period includes growing public exposure to AI chatbots. As such, there are many instances of students expressing skepticism about whether their tutor was a real person.
337
+ 4. Our strong commitment to student privacy limits access to individual-level student details, which may constrain downstream use in student modeling tasks.
 
 
338
 
339
  ## Citation [optional]
340
 
341
+ Publication in Progress. Please reach out to Matthew.Zent@eedi.co.uk for citation information.
342
 
343
+ ## Dataset Card Author
344
 
345
  Matthew Zent
346
 
347
  ## Dataset Card Contact
348
 
349
+ Matthew.Zent@eedi.co.uk