File size: 1,498 Bytes
80058a3 86de414 80058a3 86de414 bbcdd80 86de414 bbcdd80 86de414 b7d1204 86de414 83a1fbf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 | ---
license: apache-2.0
datasets:
- aripos1/gorani_dataset
language:
- ko
- en
- ja
base_model:
- unsloth/Llama-3.2-3B-Instruct-bnb-4bit
pipeline_tag: text-generation
library_name: transformers
---
# Gorani Model Card
## ์๊ฐ (Introduce)
์ด ๋ชจ๋ธ์ ๋ฒ์ญ์ ์ํ ๋ชจ๋ธ์
๋๋ค. ํ๊ตญ ๊ณ ์ ์ด์ ์ ํํ ๋ฒ์ญ์ ์์ฑํ๊ธฐ ์ํด ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด์ ์ธ์ด ๋ฐ์ดํฐ๋ฅผ ํผํฉํ์ฌ **unsloth/Llama-3.2-3B-Instruct-bnb-4bit**์ ํ์ต์์ผ ์์ฑ๋ **gorani-1B** ์
๋๋ค.
gorani๋ ํ์ฌ **ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด**๋ง ๋ฒ์ญ์ ์ง์ํฉ๋๋ค.
### ๋ชจ๋ธ ์ ๋ณด
- **๊ฐ๋ฐ์**: airpos1
- **๋ชจ๋ธ ์ ํ**: **llama**๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ํ๋ **3B** ๋งค๊ฐ๋ณ์ ๋ชจ๋ธ์ธ **gorani-3B**
- **์ง์ ์ธ์ด**: ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด
- **๋ผ์ด์ผ์ค**: **llama**
## Training Hyperparameters
- **per_device_train_batch_size**: 8
- **gradient_accumulation_steps**: 1
- **warmup_steps**: 5
- **learning_rate**: 2e-4
- **fp16**: `not is_bfloat16_supported()`
- **num_train_epochs**: 3
- **weight_decay**: 0.01
- **lr_scheduler_type**: "linear"
## ํ์ต ๋ฐ์ดํฐ
[๋ฐ์ดํฐ์
๋งํฌ](https://huggingface.co/datasets/aripos1/gorani_dataset)
## ํ์ต ์ฑ๋ฅ ๋น๊ต

## Training Results

|