File size: 825 Bytes
8376dcf
 
 
 
 
 
 
ca21f73
 
 
 
 
 
 
0be2106
2c53dc8
 
0be2106
52087f8
 
0be2106
ca21f73
0be2106
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
---
library_name: transformers
tags: []
---

# Model Card for Model ID

This is a model I accidentally trained with too low a batch size, causing the training loss to spike and essentially fail.
I found it amusing that it nevertheless does very well on EWoK, Entity Tracking, Adjective Nominalization, COMPS, and AoA. 
Maybe this says something about ourselves, how so many in society fail upwards... food for thought.


### UPDATE 
Thanks to the work of my student [Serdar Gülbahar](https://github.com/serdardoesml), the reason for this model scoring well has been traced to a few bugs in the babylm evaluation pipeline.
The issue is currently being fixed here: https://github.com/babylm/evaluation-pipeline-2025/issues/34

After this is fixed, this model should perform quite poorly, as expected.

-------------------