Update README.md
Browse files
README.md
CHANGED
|
@@ -6,7 +6,6 @@ datasets:
|
|
| 6 |
# OLMo-Bitnet-1B
|
| 7 |
|
| 8 |
OLMo-Bitnet-1B is a 1B parameter model trained using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
| 9 |
-
The result of this is that all of the parameter weights take only the values -1, 0, or 1.
|
| 10 |
|
| 11 |
It was trained on a 60B subset of the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset, so it is merely a research proof-of-concept to test out the methodolgy.
|
| 12 |
|
|
|
|
| 6 |
# OLMo-Bitnet-1B
|
| 7 |
|
| 8 |
OLMo-Bitnet-1B is a 1B parameter model trained using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
|
|
|
| 9 |
|
| 10 |
It was trained on a 60B subset of the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset, so it is merely a research proof-of-concept to test out the methodolgy.
|
| 11 |
|