Spaces:
Sleeping
feat: Add better Korean language models
Browse files**Model Changes:**
- Remove: skt/kogpt2-base-v2 (poor conversation quality)
- Add: beomi/llama-2-ko-7b (Llama 2 based Korean model, 14GB)
- Add: kyujinpy/KoT-Llama2-7B-Chat (Korean optimized Llama 2, 14GB)
- Add: nlpai-lab/kullm-polyglot-5.8b-v2 (Korea Univ NLP lab model, 12GB)
- Keep: beomi/KoAlpaca-Polyglot-5.8B (existing model, 12GB)
**Model Summary:**
- Total: 7 models (3 English, 4 Korean)
- All Korean models have memory warnings (12-14GB RAM required)
- Korean models are not suitable for HF Spaces free tier
**Documentation Updates:**
- Updated README with new model list and specifications
- Added performance notes for Korean models
- Updated examples to use new Korean models
- Clarified that Korean models require GPU tier on HF Spaces
**Important Notes:**
- Korean models provide better conversation quality
- All Korean models require high memory (12-14GB)
- Free tier users should stick to English models only
- GPU recommended for acceptable performance
π€ Generated with [Claude Code](https://claude.com/claude-code)
Co-Authored-By: Claude <noreply@anthropic.com>
|
@@ -16,7 +16,7 @@ license: mit
|
|
| 16 |
|
| 17 |
## β¨ μ£Όμ κΈ°λ₯
|
| 18 |
|
| 19 |
-
- **λ€μ€ λͺ¨λΈ μ§μ**:
|
| 20 |
- **λ‘컬 μ€ν**: Transformers λΌμ΄λΈλ¬λ¦¬λ‘ λ‘컬μμ λͺ¨λΈ μ€ν
|
| 21 |
- **API μ ν μμ**: μΈν°λ· μ°κ²° μμ΄λ μλ (첫 λ€μ΄λ‘λ ν)
|
| 22 |
- **μλ μΈμ
κ΄λ¦¬**: λͺ¨λΈ λ³κ²½ μ λν μλ μ΄κΈ°ν
|
|
@@ -30,8 +30,10 @@ license: mit
|
|
| 30 |
3. **GPT-2** - λ²μ© ν
μ€νΈ μμ± λͺ¨λΈ (~500MB)
|
| 31 |
|
| 32 |
### νκΈ λͺ¨λΈ
|
| 33 |
-
4. **
|
| 34 |
-
5. **
|
|
|
|
|
|
|
| 35 |
|
| 36 |
## π λ‘컬 μ€ν λ°©λ²
|
| 37 |
|
|
@@ -122,21 +124,23 @@ simple-chatbot-gradio/
|
|
| 122 |
## β οΈ μ νμ¬ν λ° μ£Όμμ¬ν
|
| 123 |
|
| 124 |
### μ±λ₯
|
| 125 |
-
- **CPU μ€ν**: GPU μμ΄ CPUμμ μ€νλλ―λ‘ μλ΅μ΄ λ릴 μ μμ΅λλ€ (5-
|
| 126 |
-
- **λ©λͺ¨λ¦¬**: λͺ¨λΈ ν¬κΈ°μ λ°λΌ 1-
|
| 127 |
-
- **첫 μ€ν**: λͺ¨λΈ λ€μ΄λ‘λλ‘ μκ° μμ (350MB~
|
| 128 |
|
| 129 |
### λͺ¨λΈλ³ νΉμ±
|
| 130 |
- **μμ΄ λͺ¨λΈ**: νκΈ μ
λ ₯ μ λΆμμ°μ€λ¬μ΄ μλ΅
|
| 131 |
-
- **νκΈ
|
| 132 |
-
-
|
|
|
|
| 133 |
|
| 134 |
### Hugging Face Spaces λ°°ν¬
|
| 135 |
- **λ¬΄λ£ tier**: CPU μΈμ€ν΄μ€λ§ μ 곡 (16GB RAM)
|
| 136 |
- **Space Sleep**: 48μκ° λΉνμ± μ μλ sleep, 첫 λ‘λ© λλ¦Ό
|
| 137 |
-
- **λ©λͺ¨λ¦¬ μ ν**:
|
| 138 |
-
- **첫 μ€ν**: λͺ¨λΈ λ€μ΄λ‘λλ‘ 1-
|
| 139 |
-
- **κΆμ₯ λͺ¨λΈ**: DialoGPT Small/Medium, GPT-2
|
|
|
|
| 140 |
|
| 141 |
## π§ κ°λ° λ° μ»€μ€ν°λ§μ΄μ§
|
| 142 |
|
|
|
|
| 16 |
|
| 17 |
## β¨ μ£Όμ κΈ°λ₯
|
| 18 |
|
| 19 |
+
- **λ€μ€ λͺ¨λΈ μ§μ**: 7κ° λͺ¨λΈ (μμ΄ 3κ°, νκΈ 4κ°)
|
| 20 |
- **λ‘컬 μ€ν**: Transformers λΌμ΄λΈλ¬λ¦¬λ‘ λ‘컬μμ λͺ¨λΈ μ€ν
|
| 21 |
- **API μ ν μμ**: μΈν°λ· μ°κ²° μμ΄λ μλ (첫 λ€μ΄λ‘λ ν)
|
| 22 |
- **μλ μΈμ
κ΄λ¦¬**: λͺ¨λΈ λ³κ²½ μ λν μλ μ΄κΈ°ν
|
|
|
|
| 30 |
3. **GPT-2** - λ²μ© ν
μ€νΈ μμ± λͺ¨λΈ (~500MB)
|
| 31 |
|
| 32 |
### νκΈ λͺ¨λΈ
|
| 33 |
+
4. **Llama-2-Ko 7B** - Llama 2 κΈ°λ° νκΈ λνν λͺ¨λΈ (~14GB, κ³ μ¬μ)
|
| 34 |
+
5. **KoT-Llama2-7B-Chat** - νκ΅μ΄ μ΅μ ν Llama 2 λν λͺ¨λΈ (~14GB, κ³ μ¬μ)
|
| 35 |
+
6. **KoAlpaca 5.8B** - νκΈ λνν λͺ¨λΈ (~12GB, κ³ μ¬μ)
|
| 36 |
+
7. **KULLM-Polyglot 5.8B** - κ³ λ €λ NLP μ°κ΅¬μ€ νκΈ λν λͺ¨λΈ (~12GB, κ³ μ¬μ)
|
| 37 |
|
| 38 |
## π λ‘컬 μ€ν λ°©λ²
|
| 39 |
|
|
|
|
| 124 |
## β οΈ μ νμ¬ν λ° μ£Όμμ¬ν
|
| 125 |
|
| 126 |
### μ±λ₯
|
| 127 |
+
- **CPU μ€ν**: GPU μμ΄ CPUμμ μ€νλλ―λ‘ μλ΅μ΄ λ릴 μ μμ΅λλ€ (5-30μ΄)
|
| 128 |
+
- **λ©λͺ¨λ¦¬**: λͺ¨λΈ ν¬κΈ°μ λ°λΌ 1-16GB RAM νμ
|
| 129 |
+
- **첫 μ€ν**: λͺ¨λΈ λ€μ΄λ‘λλ‘ μκ° μμ (350MB~14GB)
|
| 130 |
|
| 131 |
### λͺ¨λΈλ³ νΉμ±
|
| 132 |
- **μμ΄ λͺ¨λΈ**: νκΈ μ
λ ₯ μ λΆμμ°μ€λ¬μ΄ μλ΅
|
| 133 |
+
- **νκΈ λͺ¨λΈ (Llama 2 κΈ°λ°)**: λν νμ§ μ°μνμ§λ§ λ©λͺ¨λ¦¬ λ§μ΄ νμ (14GB+)
|
| 134 |
+
- **νκΈ λͺ¨λΈ (Polyglot κΈ°λ°)**: μ€κ° ν¬κΈ°, λν νμ§ μνΈ (12GB+)
|
| 135 |
+
- **λͺ¨λ νκΈ λͺ¨λΈ**: CPU νκ²½μμ λ§€μ° λλ¦Ό, GPU κΆμ₯
|
| 136 |
|
| 137 |
### Hugging Face Spaces λ°°ν¬
|
| 138 |
- **λ¬΄λ£ tier**: CPU μΈμ€ν΄μ€λ§ μ 곡 (16GB RAM)
|
| 139 |
- **Space Sleep**: 48μκ° λΉνμ± μ μλ sleep, 첫 λ‘λ© λλ¦Ό
|
| 140 |
+
- **λ©λͺ¨λ¦¬ μ ν**: νκΈ λͺ¨λΈλ€μ λ¬΄λ£ tierμμ μ€ν λΆκ° (12-14GB νμ)
|
| 141 |
+
- **첫 μ€ν**: λͺ¨λΈ λ€μ΄λ‘λλ‘ 1-5λΆ μμ
|
| 142 |
+
- **κΆμ₯ λͺ¨λΈ**: DialoGPT Small/Medium, GPT-2λ§ λ¬΄λ£ tierμμ μμ μ
|
| 143 |
+
- **νκΈ λν**: λ¬΄λ£ tierμμλ νκΈ λͺ¨λΈ μ¬μ© λΆκ°, μ λ£ GPU tier νμ
|
| 144 |
|
| 145 |
## π§ κ°λ° λ° μ»€μ€ν°λ§μ΄μ§
|
| 146 |
|
|
@@ -32,16 +32,29 @@ MODELS = {
|
|
| 32 |
"max_length": 80,
|
| 33 |
"language": "en",
|
| 34 |
},
|
| 35 |
-
"
|
| 36 |
-
"name": "
|
| 37 |
-
"max_length":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
"language": "ko",
|
|
|
|
| 39 |
},
|
| 40 |
"beomi/KoAlpaca-Polyglot-5.8B": {
|
| 41 |
-
"name": "KoAlpaca 5.8B (νκΈ λνν, β οΈ
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
"max_length": 150,
|
| 43 |
"language": "ko",
|
| 44 |
-
"warning": "μ΄ λͺ¨λΈμ
|
| 45 |
},
|
| 46 |
}
|
| 47 |
|
|
@@ -225,8 +238,8 @@ with gr.Blocks(
|
|
| 225 |
examples=[
|
| 226 |
["Hello! How are you?", "microsoft/DialoGPT-small"],
|
| 227 |
["Tell me a joke", "microsoft/DialoGPT-medium"],
|
| 228 |
-
["μλ
νμΈμ! μ€λ λ μ¨κ°
|
| 229 |
-
["μΈκ³΅μ§λ₯μ λν΄ μ€λͺ
ν΄μ£ΌμΈμ.", "
|
| 230 |
],
|
| 231 |
)
|
| 232 |
|
|
|
|
| 32 |
"max_length": 80,
|
| 33 |
"language": "en",
|
| 34 |
},
|
| 35 |
+
"beomi/llama-2-ko-7b": {
|
| 36 |
+
"name": "Llama-2-Ko 7B (νκΈ λνν, β οΈ 14GB+ RAM νμ)",
|
| 37 |
+
"max_length": 150,
|
| 38 |
+
"language": "ko",
|
| 39 |
+
"warning": "μ΄ λͺ¨λΈμ 14GB μ΄μμ λ©λͺ¨λ¦¬κ° νμν©λλ€. HF Spaces λ¬΄λ£ tierμμλ λ©λͺ¨λ¦¬ λΆμ‘±μΌλ‘ μ€νλμ§ μμ μ μμ΅λλ€.",
|
| 40 |
+
},
|
| 41 |
+
"kyujinpy/KoT-Llama2-7B-Chat": {
|
| 42 |
+
"name": "KoT-Llama2-7B-Chat (νκΈ λν, β οΈ 14GB+ RAM νμ)",
|
| 43 |
+
"max_length": 150,
|
| 44 |
"language": "ko",
|
| 45 |
+
"warning": "μ΄ λͺ¨λΈμ 14GB μ΄μμ λ©λͺ¨λ¦¬κ° νμν©λλ€. HF Spaces λ¬΄λ£ tierμμλ λ©λͺ¨λ¦¬ λΆμ‘±μΌλ‘ μ€νλμ§ μμ μ μμ΅λλ€.",
|
| 46 |
},
|
| 47 |
"beomi/KoAlpaca-Polyglot-5.8B": {
|
| 48 |
+
"name": "KoAlpaca 5.8B (νκΈ λνν, β οΈ 12GB+ RAM νμ)",
|
| 49 |
+
"max_length": 150,
|
| 50 |
+
"language": "ko",
|
| 51 |
+
"warning": "μ΄ λͺ¨λΈμ 12GB μ΄μμ λ©λͺ¨λ¦¬κ° νμν©λλ€. HF Spaces λ¬΄λ£ tierμμλ λ©λͺ¨λ¦¬ λΆμ‘±μΌλ‘ μ€νλμ§ μμ μ μμ΅λλ€.",
|
| 52 |
+
},
|
| 53 |
+
"nlpai-lab/kullm-polyglot-5.8b-v2": {
|
| 54 |
+
"name": "KULLM-Polyglot 5.8B (νκΈ λν, β οΈ 12GB+ RAM νμ)",
|
| 55 |
"max_length": 150,
|
| 56 |
"language": "ko",
|
| 57 |
+
"warning": "μ΄ λͺ¨λΈμ 12GB μ΄μμ λ©λͺ¨λ¦¬κ° νμν©λλ€. HF Spaces λ¬΄λ£ tierμμλ λ©λͺ¨λ¦¬ λΆμ‘±μΌλ‘ μ€νλμ§ μμ μ μμ΅λλ€.",
|
| 58 |
},
|
| 59 |
}
|
| 60 |
|
|
|
|
| 238 |
examples=[
|
| 239 |
["Hello! How are you?", "microsoft/DialoGPT-small"],
|
| 240 |
["Tell me a joke", "microsoft/DialoGPT-medium"],
|
| 241 |
+
["μλ
νμΈμ! μ€λ λ μ¨κ° μ΄λμ?", "beomi/llama-2-ko-7b"],
|
| 242 |
+
["μΈκ³΅μ§λ₯μ λν΄ κ°λ¨ν μ€λͺ
ν΄μ£ΌμΈμ.", "kyujinpy/KoT-Llama2-7B-Chat"],
|
| 243 |
],
|
| 244 |
)
|
| 245 |
|