alex4cip Claude commited on
Commit
c6d4144
Β·
1 Parent(s): fb55a82

feat: Add better Korean language models

Browse files

**Model Changes:**
- Remove: skt/kogpt2-base-v2 (poor conversation quality)
- Add: beomi/llama-2-ko-7b (Llama 2 based Korean model, 14GB)
- Add: kyujinpy/KoT-Llama2-7B-Chat (Korean optimized Llama 2, 14GB)
- Add: nlpai-lab/kullm-polyglot-5.8b-v2 (Korea Univ NLP lab model, 12GB)
- Keep: beomi/KoAlpaca-Polyglot-5.8B (existing model, 12GB)

**Model Summary:**
- Total: 7 models (3 English, 4 Korean)
- All Korean models have memory warnings (12-14GB RAM required)
- Korean models are not suitable for HF Spaces free tier

**Documentation Updates:**
- Updated README with new model list and specifications
- Added performance notes for Korean models
- Updated examples to use new Korean models
- Clarified that Korean models require GPU tier on HF Spaces

**Important Notes:**
- Korean models provide better conversation quality
- All Korean models require high memory (12-14GB)
- Free tier users should stick to English models only
- GPU recommended for acceptable performance

πŸ€– Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

Files changed (2) hide show
  1. README.md +15 -11
  2. app.py +20 -7
README.md CHANGED
@@ -16,7 +16,7 @@ license: mit
16
 
17
  ## ✨ μ£Όμš” κΈ°λŠ₯
18
 
19
- - **닀쀑 λͺ¨λΈ 지원**: 5개 λͺ¨λΈ (μ˜μ–΄ 3개, ν•œκΈ€ 2개)
20
  - **둜컬 μ‹€ν–‰**: Transformers 라이브러리둜 λ‘œμ»¬μ—μ„œ λͺ¨λΈ μ‹€ν–‰
21
  - **API μ œν•œ μ—†μŒ**: 인터넷 μ—°κ²° 없이도 μž‘λ™ (첫 λ‹€μš΄λ‘œλ“œ ν›„)
22
  - **μžλ™ μ„Έμ…˜ 관리**: λͺ¨λΈ λ³€κ²½ μ‹œ λŒ€ν™” μžλ™ μ΄ˆκΈ°ν™”
@@ -30,8 +30,10 @@ license: mit
30
  3. **GPT-2** - λ²”μš© ν…μŠ€νŠΈ 생성 λͺ¨λΈ (~500MB)
31
 
32
  ### ν•œκΈ€ λͺ¨λΈ
33
- 4. **KoGPT-2** - SKT의 ν•œκΈ€ νŠΉν™” λͺ¨λΈ (~500MB)
34
- 5. **KoAlpaca 5.8B** - λŒ€ν™”ν˜• ν•œκΈ€ λͺ¨λΈ, 고사양 ν•„μš” (~12GB)
 
 
35
 
36
  ## πŸš€ 둜컬 μ‹€ν–‰ 방법
37
 
@@ -122,21 +124,23 @@ simple-chatbot-gradio/
122
  ## ⚠️ μ œν•œμ‚¬ν•­ 및 μ£Όμ˜μ‚¬ν•­
123
 
124
  ### μ„±λŠ₯
125
- - **CPU μ‹€ν–‰**: GPU 없이 CPUμ—μ„œ μ‹€ν–‰λ˜λ―€λ‘œ 응닡이 느릴 수 μžˆμŠ΅λ‹ˆλ‹€ (5-10초)
126
- - **λ©”λͺ¨λ¦¬**: λͺ¨λΈ 크기에 따라 1-8GB RAM ν•„μš”
127
- - **첫 μ‹€ν–‰**: λͺ¨λΈ λ‹€μš΄λ‘œλ“œλ‘œ μ‹œκ°„ μ†Œμš” (350MB~12GB)
128
 
129
  ### λͺ¨λΈλ³„ νŠΉμ„±
130
  - **μ˜μ–΄ λͺ¨λΈ**: ν•œκΈ€ μž…λ ₯ μ‹œ λΆ€μžμ—°μŠ€λŸ¬μš΄ 응닡
131
- - **ν•œκΈ€ λͺ¨λΈ**: μ˜μ–΄ μž…λ ₯ μ‹œ μ„±λŠ₯ μ €ν•˜
132
- - **KoAlpaca 5.8B**: 8GB+ RAM ν•„μš”, CPUμ—μ„œ 맀우 느림
 
133
 
134
  ### Hugging Face Spaces 배포
135
  - **무료 tier**: CPU μΈμŠ€ν„΄μŠ€λ§Œ 제곡 (16GB RAM)
136
  - **Space Sleep**: 48μ‹œκ°„ λΉ„ν™œμ„± μ‹œ μžλ™ sleep, 첫 λ‘œλ”© 느림
137
- - **λ©”λͺ¨λ¦¬ μ œν•œ**: KoAlpaca 5.8BλŠ” 무료 tierμ—μ„œ μ‹€ν–‰ λΆˆκ°€ (8GB+ ν•„μš”)
138
- - **첫 μ‹€ν–‰**: λͺ¨λΈ λ‹€μš΄λ‘œλ“œλ‘œ 1-3λΆ„ μ†Œμš”
139
- - **ꢌμž₯ λͺ¨λΈ**: DialoGPT Small/Medium, GPT-2, KoGPT-2 (무료 tierμ—μ„œ μ•ˆμ •μ )
 
140
 
141
  ## πŸ”§ 개발 및 μ»€μŠ€ν„°λ§ˆμ΄μ§•
142
 
 
16
 
17
  ## ✨ μ£Όμš” κΈ°λŠ₯
18
 
19
+ - **닀쀑 λͺ¨λΈ 지원**: 7개 λͺ¨λΈ (μ˜μ–΄ 3개, ν•œκΈ€ 4개)
20
  - **둜컬 μ‹€ν–‰**: Transformers 라이브러리둜 λ‘œμ»¬μ—μ„œ λͺ¨λΈ μ‹€ν–‰
21
  - **API μ œν•œ μ—†μŒ**: 인터넷 μ—°κ²° 없이도 μž‘λ™ (첫 λ‹€μš΄λ‘œλ“œ ν›„)
22
  - **μžλ™ μ„Έμ…˜ 관리**: λͺ¨λΈ λ³€κ²½ μ‹œ λŒ€ν™” μžλ™ μ΄ˆκΈ°ν™”
 
30
  3. **GPT-2** - λ²”μš© ν…μŠ€νŠΈ 생성 λͺ¨λΈ (~500MB)
31
 
32
  ### ν•œκΈ€ λͺ¨λΈ
33
+ 4. **Llama-2-Ko 7B** - Llama 2 기반 ν•œκΈ€ λŒ€ν™”ν˜• λͺ¨λΈ (~14GB, 고사양)
34
+ 5. **KoT-Llama2-7B-Chat** - ν•œκ΅­μ–΄ μ΅œμ ν™” Llama 2 λŒ€ν™” λͺ¨λΈ (~14GB, 고사양)
35
+ 6. **KoAlpaca 5.8B** - ν•œκΈ€ λŒ€ν™”ν˜• λͺ¨λΈ (~12GB, 고사양)
36
+ 7. **KULLM-Polyglot 5.8B** - κ³ λ €λŒ€ NLP 연ꡬ싀 ν•œκΈ€ λŒ€ν™” λͺ¨λΈ (~12GB, 고사양)
37
 
38
  ## πŸš€ 둜컬 μ‹€ν–‰ 방법
39
 
 
124
  ## ⚠️ μ œν•œμ‚¬ν•­ 및 μ£Όμ˜μ‚¬ν•­
125
 
126
  ### μ„±λŠ₯
127
+ - **CPU μ‹€ν–‰**: GPU 없이 CPUμ—μ„œ μ‹€ν–‰λ˜λ―€λ‘œ 응닡이 느릴 수 μžˆμŠ΅λ‹ˆλ‹€ (5-30초)
128
+ - **λ©”λͺ¨λ¦¬**: λͺ¨λΈ 크기에 따라 1-16GB RAM ν•„μš”
129
+ - **첫 μ‹€ν–‰**: λͺ¨λΈ λ‹€μš΄λ‘œλ“œλ‘œ μ‹œκ°„ μ†Œμš” (350MB~14GB)
130
 
131
  ### λͺ¨λΈλ³„ νŠΉμ„±
132
  - **μ˜μ–΄ λͺ¨λΈ**: ν•œκΈ€ μž…λ ₯ μ‹œ λΆ€μžμ—°μŠ€λŸ¬μš΄ 응닡
133
+ - **ν•œκΈ€ λͺ¨λΈ (Llama 2 기반)**: λŒ€ν™” ν’ˆμ§ˆ μš°μˆ˜ν•˜μ§€λ§Œ λ©”λͺ¨λ¦¬ 많이 ν•„μš” (14GB+)
134
+ - **ν•œκΈ€ λͺ¨λΈ (Polyglot 기반)**: 쀑간 크기, λŒ€ν™” ν’ˆμ§ˆ μ–‘ν˜Έ (12GB+)
135
+ - **λͺ¨λ“  ν•œκΈ€ λͺ¨λΈ**: CPU ν™˜κ²½μ—μ„œ 맀우 느림, GPU ꢌμž₯
136
 
137
  ### Hugging Face Spaces 배포
138
  - **무료 tier**: CPU μΈμŠ€ν„΄μŠ€λ§Œ 제곡 (16GB RAM)
139
  - **Space Sleep**: 48μ‹œκ°„ λΉ„ν™œμ„± μ‹œ μžλ™ sleep, 첫 λ‘œλ”© 느림
140
+ - **λ©”λͺ¨λ¦¬ μ œν•œ**: ν•œκΈ€ λͺ¨λΈλ“€μ€ 무료 tierμ—μ„œ μ‹€ν–‰ λΆˆκ°€ (12-14GB ν•„μš”)
141
+ - **첫 μ‹€ν–‰**: λͺ¨λΈ λ‹€μš΄λ‘œλ“œλ‘œ 1-5λΆ„ μ†Œμš”
142
+ - **ꢌμž₯ λͺ¨λΈ**: DialoGPT Small/Medium, GPT-2만 무료 tierμ—μ„œ μ•ˆμ •μ 
143
+ - **ν•œκΈ€ λŒ€ν™”**: 무료 tierμ—μ„œλŠ” ν•œκΈ€ λͺ¨λΈ μ‚¬μš© λΆˆκ°€, 유료 GPU tier ν•„μš”
144
 
145
  ## πŸ”§ 개발 및 μ»€μŠ€ν„°λ§ˆμ΄μ§•
146
 
app.py CHANGED
@@ -32,16 +32,29 @@ MODELS = {
32
  "max_length": 80,
33
  "language": "en",
34
  },
35
- "skt/kogpt2-base-v2": {
36
- "name": "KoGPT-2 (ν•œκΈ€ νŠΉν™”)",
37
- "max_length": 100,
 
 
 
 
 
 
38
  "language": "ko",
 
39
  },
40
  "beomi/KoAlpaca-Polyglot-5.8B": {
41
- "name": "KoAlpaca 5.8B (ν•œκΈ€ λŒ€ν™”ν˜•, ⚠️ 8GB+ RAM ν•„μš”)",
 
 
 
 
 
 
42
  "max_length": 150,
43
  "language": "ko",
44
- "warning": "이 λͺ¨λΈμ€ 8GB μ΄μƒμ˜ λ©”λͺ¨λ¦¬κ°€ ν•„μš”ν•©λ‹ˆλ‹€. HF Spaces 무료 tierμ—μ„œλŠ” λ©”λͺ¨λ¦¬ λΆ€μ‘±μœΌλ‘œ μ‹€ν–‰λ˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€.",
45
  },
46
  }
47
 
@@ -225,8 +238,8 @@ with gr.Blocks(
225
  examples=[
226
  ["Hello! How are you?", "microsoft/DialoGPT-small"],
227
  ["Tell me a joke", "microsoft/DialoGPT-medium"],
228
- ["μ•ˆλ…•ν•˜μ„Έμš”! 였늘 날씨가 μ’‹λ„€μš”.", "skt/kogpt2-base-v2"],
229
- ["인곡지λŠ₯에 λŒ€ν•΄ μ„€λͺ…ν•΄μ£Όμ„Έμš”.", "skt/kogpt2-base-v2"],
230
  ],
231
  )
232
 
 
32
  "max_length": 80,
33
  "language": "en",
34
  },
35
+ "beomi/llama-2-ko-7b": {
36
+ "name": "Llama-2-Ko 7B (ν•œκΈ€ λŒ€ν™”ν˜•, ⚠️ 14GB+ RAM ν•„μš”)",
37
+ "max_length": 150,
38
+ "language": "ko",
39
+ "warning": "이 λͺ¨λΈμ€ 14GB μ΄μƒμ˜ λ©”λͺ¨λ¦¬κ°€ ν•„μš”ν•©λ‹ˆλ‹€. HF Spaces 무료 tierμ—μ„œλŠ” λ©”λͺ¨λ¦¬ λΆ€μ‘±μœΌλ‘œ μ‹€ν–‰λ˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€.",
40
+ },
41
+ "kyujinpy/KoT-Llama2-7B-Chat": {
42
+ "name": "KoT-Llama2-7B-Chat (ν•œκΈ€ λŒ€ν™”, ⚠️ 14GB+ RAM ν•„μš”)",
43
+ "max_length": 150,
44
  "language": "ko",
45
+ "warning": "이 λͺ¨λΈμ€ 14GB μ΄μƒμ˜ λ©”λͺ¨λ¦¬κ°€ ν•„μš”ν•©λ‹ˆλ‹€. HF Spaces 무료 tierμ—μ„œλŠ” λ©”λͺ¨λ¦¬ λΆ€μ‘±μœΌλ‘œ μ‹€ν–‰λ˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€.",
46
  },
47
  "beomi/KoAlpaca-Polyglot-5.8B": {
48
+ "name": "KoAlpaca 5.8B (ν•œκΈ€ λŒ€ν™”ν˜•, ⚠️ 12GB+ RAM ν•„μš”)",
49
+ "max_length": 150,
50
+ "language": "ko",
51
+ "warning": "이 λͺ¨λΈμ€ 12GB μ΄μƒμ˜ λ©”λͺ¨λ¦¬κ°€ ν•„μš”ν•©λ‹ˆλ‹€. HF Spaces 무료 tierμ—μ„œλŠ” λ©”λͺ¨λ¦¬ λΆ€μ‘±μœΌλ‘œ μ‹€ν–‰λ˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€.",
52
+ },
53
+ "nlpai-lab/kullm-polyglot-5.8b-v2": {
54
+ "name": "KULLM-Polyglot 5.8B (ν•œκΈ€ λŒ€ν™”, ⚠️ 12GB+ RAM ν•„μš”)",
55
  "max_length": 150,
56
  "language": "ko",
57
+ "warning": "이 λͺ¨λΈμ€ 12GB μ΄μƒμ˜ λ©”λͺ¨λ¦¬κ°€ ν•„μš”ν•©λ‹ˆλ‹€. HF Spaces 무료 tierμ—μ„œλŠ” λ©”λͺ¨λ¦¬ λΆ€μ‘±μœΌλ‘œ μ‹€ν–‰λ˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€.",
58
  },
59
  }
60
 
 
238
  examples=[
239
  ["Hello! How are you?", "microsoft/DialoGPT-small"],
240
  ["Tell me a joke", "microsoft/DialoGPT-medium"],
241
+ ["μ•ˆλ…•ν•˜μ„Έμš”! 였늘 날씨가 μ–΄λ•Œμš”?", "beomi/llama-2-ko-7b"],
242
+ ["인곡지λŠ₯에 λŒ€ν•΄ κ°„λ‹¨νžˆ μ„€λͺ…ν•΄μ£Όμ„Έμš”.", "kyujinpy/KoT-Llama2-7B-Chat"],
243
  ],
244
  )
245