phueb commited on
Commit
151a52a
·
1 Parent(s): 27cc6d7

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -0
README.md ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## BabyBERTA
2
+
3
+ ### Overview
4
+
5
+ BabyBERTa is a light-weight version of RoBERTa trained on 5M words of American-English child-directed input.
6
+ It is intended for language acquisition research, on a single desktop with a single GPU - no high-performance computing infrastructure needed.
7
+
8
+ The three provided models are randomly selected from 10 that were trained and reported in the paper.
9
+
10
+ ## Loading the tokenizer
11
+
12
+ BabyBERTa was trained with `add_prefix_space=True`, so it will not work properly with the tokenizer defaults.
13
+ Make sure to load the tokenizer as follows:
14
+
15
+ ```python
16
+ tokenizer = RobertaTokenizerFast.from_pretrained("phueb/BabyBERTa",
17
+ add_prefix_space=True)
18
+ ```
19
+
20
+ ### Hyper-Parameters
21
+
22
+ See the paper for details.
23
+ All provided models were trained for 400K steps with a batch size of 16.
24
+ Importantly, BabyBERTa never predicts unmasked tokens during training - `unmask_prob` is set to zero.
25
+
26
+
27
+ ### Performance
28
+
29
+ BabyBerta was developed for learning grammatical knowledge from child-directed input.
30
+ Its grammatical knowledge was evaluated using the [Zorro](https://github.com/phueb/Zorro) test suite.
31
+ The best model achieves an overall accuracy of 80.3,
32
+ comparable to RoBERTa-base, which achieves an overall accuracy of 82.6 on the latest version of Zorro (as of October, 2021).
33
+ Both values differ slightly from those reported in the paper (Huebner et al., 2020).
34
+ There are two reasons for this:
35
+ 1. Performance of RoBERTa-base is slightly larger because the authors previously lower-cased all words in Zorro before evaluation.
36
+ Lower-casing of proper nouns is detrimental to RoBERTa-base because RoBERTa-base has likely been trained on proper nouns that are primarily title-cased.
37
+ In contrast, because BabyBERTa is not case-sensitive, its performance is not influenced by this change.
38
+ 2. The latest version of Zorro no longer contains ambiguous content words such as "Spanish" which can be both a noun and an adjective.
39
+ this resulted in a small reduction in the performance of BabyBERTa.
40
+
41
+ | Model Name | Accuracy (holistic scoring) | Accuracy (MLM-scoring) |
42
+ |----------------------------------------|------------------------------|------------|
43
+ | [BabyBERTa-1][link-BabyBERTa-1] | 80.3 | 79.9 |
44
+ | [BabyBERTa-2][link-BabyBERTa-2] | 80.3 | 79.9 |
45
+ | [BabyBERTa-3][link-BabyBERTa-3] | 80.3 | 79.9 |
46
+
47
+
48
+
49
+ ### Additional Information
50
+
51
+ This model was trained by [Philip Huebner](https://philhuebner.com), currently at the [UIUC Language and Learning Lab](http://www.learninglanguagelab.org).
52
+
53
+ More info can be found [here](https://github.com/phueb/BabyBERTa).
54
+
55
+
56
+ [link-BabyBERTa-1]: https://huggingface.co/phueb/BabyBERTa-1
57
+ [link-BabyBERTa-2]: https://huggingface.co/phueb/BabyBERTa-2
58
+ [link-BabyBERTa-3]: https://huggingface.co/phueb/BabyBERTa-3
59
+
60
+ ---
61
+ language:
62
+ - en
63
+ tags:
64
+ - child-directed-language
65
+ - acquisition
66
+ ---