Update README.md
Browse files
README.md
CHANGED
|
@@ -27,7 +27,6 @@ Despite being trained on only 72 billion tokens of text, the models outperform m
|
|
| 27 |
# Benchmarks
|
| 28 |
|
| 29 |
**Pretrained (Temperature 0)**
|
| 30 |
-
|
| 31 |
|**Benchmark**|**TinyCodeLM 150M** |**TinyCodeLM 400M** |
|
| 32 |
| :--------------------- | -----------------: | -----------------: |
|
| 33 |
| HumanEval, pass@1 | 6.1 | 6.7 |
|
|
@@ -35,7 +34,6 @@ Despite being trained on only 72 billion tokens of text, the models outperform m
|
|
| 35 |
|
| 36 |
|
| 37 |
**Edit Sequence / Instruction Tuned (Temperature-Tuned)**
|
| 38 |
-
|
| 39 |
|**Benchmark** |**TinyCodeLM 150M** |**TinyCodeLM 400M** |
|
| 40 |
| :----------- | -----------------: | -----------------: |
|
| 41 |
| HumanEval, pass@1 | 12.8 | 13.4 |
|
|
@@ -47,7 +45,7 @@ Despite being trained on only 72 billion tokens of text, the models outperform m
|
|
| 47 |
# Citation
|
| 48 |
|
| 49 |
```
|
| 50 |
-
@misc{
|
| 51 |
title={Training Language Models on Synthetic Edit Sequences Improves Code Synthesis},
|
| 52 |
author={Ulyana Piterbarg and Lerrel Pinto and Rob Fergus},
|
| 53 |
year={2024},
|
|
|
|
| 27 |
# Benchmarks
|
| 28 |
|
| 29 |
**Pretrained (Temperature 0)**
|
|
|
|
| 30 |
|**Benchmark**|**TinyCodeLM 150M** |**TinyCodeLM 400M** |
|
| 31 |
| :--------------------- | -----------------: | -----------------: |
|
| 32 |
| HumanEval, pass@1 | 6.1 | 6.7 |
|
|
|
|
| 34 |
|
| 35 |
|
| 36 |
**Edit Sequence / Instruction Tuned (Temperature-Tuned)**
|
|
|
|
| 37 |
|**Benchmark** |**TinyCodeLM 150M** |**TinyCodeLM 400M** |
|
| 38 |
| :----------- | -----------------: | -----------------: |
|
| 39 |
| HumanEval, pass@1 | 12.8 | 13.4 |
|
|
|
|
| 45 |
# Citation
|
| 46 |
|
| 47 |
```
|
| 48 |
+
@misc{piterbarg2024editseq,
|
| 49 |
title={Training Language Models on Synthetic Edit Sequences Improves Code Synthesis},
|
| 50 |
author={Ulyana Piterbarg and Lerrel Pinto and Rob Fergus},
|
| 51 |
year={2024},
|