File size: 825 Bytes
8376dcf ca21f73 0be2106 2c53dc8 0be2106 52087f8 0be2106 ca21f73 0be2106 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
This is a model I accidentally trained with too low a batch size, causing the training loss to spike and essentially fail.
I found it amusing that it nevertheless does very well on EWoK, Entity Tracking, Adjective Nominalization, COMPS, and AoA.
Maybe this says something about ourselves, how so many in society fail upwards... food for thought.
### UPDATE
Thanks to the work of my student [Serdar Gülbahar](https://github.com/serdardoesml), the reason for this model scoring well has been traced to a few bugs in the babylm evaluation pipeline.
The issue is currently being fixed here: https://github.com/babylm/evaluation-pipeline-2025/issues/34
After this is fixed, this model should perform quite poorly, as expected.
-------------------
|