Update README.md
Browse files
README.md
CHANGED
|
@@ -4,65 +4,40 @@ datasets:
|
|
| 4 |
- aripos1/gorani_dataset
|
| 5 |
language:
|
| 6 |
- ko
|
|
|
|
|
|
|
| 7 |
base_model:
|
| 8 |
- unsloth/Llama-3.2-3B-Instruct-bnb-4bit
|
| 9 |
pipeline_tag: text-generation
|
| 10 |
library_name: transformers
|
| 11 |
---
|
| 12 |
-
# Gorani
|
| 13 |
-
|
| 14 |
-
##
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
-
|
| 20 |
-
- **
|
| 21 |
-
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
Dataset: Custom dataset (Korean-English parallel corpus)
|
| 43 |
-
Frameworks: Hugging Face transformers, peft
|
| 44 |
-
Training Hardware: NVIDIA A100 40GB
|
| 45 |
-
Training Time: 24 hours
|
| 46 |
-
π License
|
| 47 |
-
This model is licensed under the Apache 2.0 License.
|
| 48 |
-
|
| 49 |
-
---
|
| 50 |
-
|
| 51 |
-
## β
**2οΈβ£ Model Card μμ± ν μ
λ‘λ λ°©λ²**
|
| 52 |
-
Model Cardλ `README.md` νμΌμ Hugging Face μ μ₯μμ μ
λ‘λνλ©΄ μλμΌλ‘ λ°μλ¨.
|
| 53 |
-
|
| 54 |
-
### **πΉ λ°©λ² 1: Python μ½λλ‘ μ
λ‘λ**
|
| 55 |
-
```python
|
| 56 |
-
from huggingface_hub import HfApi
|
| 57 |
-
|
| 58 |
-
api = HfApi()
|
| 59 |
-
repo_id = "aripos1/gorani-lora-merged"
|
| 60 |
-
|
| 61 |
-
# README.md νμΌ μ
λ‘λ
|
| 62 |
-
api.upload_file(
|
| 63 |
-
path_or_fileobj="README.md",
|
| 64 |
-
path_in_repo="README.md",
|
| 65 |
-
repo_id=repo_id
|
| 66 |
-
)
|
| 67 |
-
|
| 68 |
-
print("β
Model Card (README.md) μ
λ‘λ μλ£!")
|
|
|
|
| 4 |
- aripos1/gorani_dataset
|
| 5 |
language:
|
| 6 |
- ko
|
| 7 |
+
- en
|
| 8 |
+
- ja
|
| 9 |
base_model:
|
| 10 |
- unsloth/Llama-3.2-3B-Instruct-bnb-4bit
|
| 11 |
pipeline_tag: text-generation
|
| 12 |
library_name: transformers
|
| 13 |
---
|
| 14 |
+
# Gorani Model Card
|
| 15 |
+
|
| 16 |
+
## μκ° (Introduce)
|
| 17 |
+
μ΄ λͺ¨λΈμ λ²μμ μν λͺ¨λΈμ
λλ€. νκ΅ κ³ μ μ΄μ μ νν λ²μμ μμ±νκΈ° μν΄ νκ΅μ΄, μμ΄, μΌλ³Έμ΄μ μΈμ΄ λ°μ΄ν°λ₯Ό νΌν©νμ¬ **unsloth/Llama-3.2-1B-Instruct-bnb-4bit**μ νμ΅μμΌ μμ±λ **gorani-1B** μ
λλ€.
|
| 18 |
+
goraniλ νμ¬ **νκ΅μ΄, μμ΄, μΌλ³Έμ΄**λ§ λ²μμ μ§μν©λλ€.
|
| 19 |
+
|
| 20 |
+
### λͺ¨λΈ μ 보
|
| 21 |
+
- **κ°λ°μ**: haeun0420
|
| 22 |
+
- **λͺ¨λΈ μ ν**: **llama**λ₯Ό κΈ°λ°μΌλ‘ νλ **1B** λ§€κ°λ³μ λͺ¨λΈμΈ **gorani-1B**
|
| 23 |
+
- **μ§μ μΈμ΄**: νκ΅μ΄, μμ΄, μΌλ³Έμ΄
|
| 24 |
+
- **λΌμ΄μΌμ€**: **llama**
|
| 25 |
+
|
| 26 |
+
## Training Hyperparameters
|
| 27 |
+
- **per_device_train_batch_size**: 8
|
| 28 |
+
- **gradient_accumulation_steps**: 1
|
| 29 |
+
- **warmup_steps**: 5
|
| 30 |
+
- **learning_rate**: 2e-4
|
| 31 |
+
- **fp16**: `not is_bfloat16_supported()`
|
| 32 |
+
- **num_train_epochs**: 3
|
| 33 |
+
- **weight_decay**: 0.01
|
| 34 |
+
- **lr_scheduler_type**: "linear"
|
| 35 |
+
|
| 36 |
+
## νμ΅ λ°μ΄ν°
|
| 37 |
+
[λ°μ΄ν°μ
λ§ν¬](https://huggingface.co/datasets/aripos1/gorani_dataset)
|
| 38 |
+
|
| 39 |
+
## νμ΅ μ±λ₯ λΉκ΅
|
| 40 |
+
![μ μ λΉκ΅ κ·Έλν]
|
| 41 |
+
|
| 42 |
+
## Training Results
|
| 43 |
+
![Training Loss κ·Έλν- μ€νμμ νμ΅ μ§ν κ·Έλν]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|