Original Model: LiquidAI/LFM2-8B-A1B
Base ๋ชจ๋ธ์ ๋๋ค.
- REAP์ผ๋ก 6B-A1B๋ก ๋ณ๊ฒฝ
- ์ฌ์ฉ ๋ฐ์ดํฐ: allenai/tulu-3-sft-personas-math
- ๋ค๊ตญ์ด ๋ฅ๋ ฅ/์ฝ๋ฉ ๋ฅ๋ ฅ์ ์ ๊ฑฐํ๊ณ ๋๋จธ์ง ๋ฅ๋ ฅ์ ๋จ๊ธฐ๊ธฐ ์ํด ๋ค์ํ ๋ฐ์ดํฐ ์กฐํฉ์ ์คํํ์ง๋ง, ์ํ ๊ณ์ด ๋ฐ์ดํฐ๋ฅผ ์ฐ๋ ๊ฒ ์ ์ผ ๋ณด์กด๋๋ ํ๊ท ์ ์๊ฐ ๋์์.
- LowRankClone์ผ๋ก 3B-A0.6B๋ก ๋ณ๊ฒฝ
- ์ฌ์ฉ ๋ฐ์ดํฐ: ์ฝ 10B ํ ํฐ
- 70% HuggingFaceTB/smollm-corpus/fineweb-edu-dedup (score 3.5 ์ด์)
- 25% C10X/finepdfs-edu-hq
- 5% OpenHermes-2.5
- ์๋ ๋ ผ๋ฌธ์์๋ Fineweb-edu + OpenHermes ์ฌ์ฉ
- ์ฌ์ฉ ๋ฐ์ดํฐ: ์ฝ 10B ํ ํฐ
5-shot ๋ฒค์น(lm-evaluation-harness, loglikelihood test)
- ARC-C, HellaSwag๋
acc_norm์ฌ์ฉ
| Name | Param | Active | MMLU | GPQA_main | PIQA | ARC-C | HellaSwag |
|---|---|---|---|---|---|---|---|
| LiquidAI/LFM2-8B-A1B | 8.3B | 1.2B | 64.84 | 25.89 | 76.44 | 60.75 | 73.34 |
| LiquidAI/LFM2-2.6B | 2.6B | 2.6B | 64.62 | 35.04 | 77.53 | 56.14 | 72.37 |
| LiquidAI/LFM2-1.2B | 1.2B | 1.2B | 55.12 | 30.36 | 73.50 | 56.06 | 63.46 |
| LiquidAI/LFM2-700M | 0.7B | 0.7B | 49.42 | 31.03 | 72.25 | 51.19 | 58.02 |
| werty1248/LFM2-6B-A1B-REAP | 6.3B | 1.2B | 56.43 | 32.37 | 71.27 | 52.73 | 65.33 |
| werty1248/LRCLFM2MoE-base-checkpoint-160000 | 3.2B | 0.6B | 44.49 | 29.69 | 71.11 | 49.40 | 54.77 |
๊ตํ: REAP๋ ๋ชจ๋ ํ์ต์ด ๋๋ ๋ค์์ ํ์(distillation์ด ๋์๊ฒ ๋จ)
์ด ๋น์ฉ: 2x B200 ์์ 110์๊ฐ (~$600)
- Downloads last month
- 50