Commit
·
4a0ff60
1
Parent(s):
b869cdf
Update README.md
Browse files
README.md
CHANGED
|
@@ -11,7 +11,10 @@ tags:
|
|
| 11 |
|
| 12 |
# CrystalCoder
|
| 13 |
|
| 14 |
-
|
|
|
|
|
|
|
|
|
|
| 15 |
This model excels in balancing natural language processing and coding capabilities.
|
| 16 |
Despite being trained on a smaller dataset of 1.4 trillion tokens—compared to LLaMA 2's 2 trillion—CrystalCoder surpasses LLaMA 2 in some challenging English and coding tasks.
|
| 17 |
It demonstrates superior performance in benchmarks like MMLU, HumanEval, and MBPP.
|
|
|
|
| 11 |
|
| 12 |
# CrystalCoder
|
| 13 |
|
| 14 |
+
<center><img src="crystalcoder_logo.jpg" alt="crystal coder logo" width="300"/></center>
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
CrystalCoder is a 7B parameter language model, distinctively trained on the SlimPajama and StarCoder datasets.
|
| 18 |
This model excels in balancing natural language processing and coding capabilities.
|
| 19 |
Despite being trained on a smaller dataset of 1.4 trillion tokens—compared to LLaMA 2's 2 trillion—CrystalCoder surpasses LLaMA 2 in some challenging English and coding tasks.
|
| 20 |
It demonstrates superior performance in benchmarks like MMLU, HumanEval, and MBPP.
|