File size: 1,686 Bytes
94dfac1 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 | ---
license: mit
datasets:
- maywell/koVast
language:
- ko
base_model:
- blueapple8259/tzem
pipeline_tag: question-answering
library_name: transformers
---
[tzem](https://huggingface.co/blueapple8259/tzem) ๋ชจ๋ธ์ instruct ๋ฐ์ดํฐ๋ก ํ์ธํ๋ ํ ๋ชจ๋ธ์
๋๋ค.
## ํ๋กฌํํธ ํ
ํ๋ฆฟ
```
**์ฌ์ฉ์:** {prompt}
**์ธ๊ณต์ง๋ฅ:**
```
## ์ฌ์ฉ
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "blueapple8259/tzem-instruct"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
prompt = "์ฌ๊ธฐ์ ํ๋กฌํํธ ์
๋ ฅ"
text = f"**์ฌ์ฉ์:** {prompt}\n**์ธ๊ณต์ง๋ฅ:"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=100,
pad_token_id=tokenizer.pad_token_id,
eos_token_id=tokenizer.eos_token_id,
do_sample=True,
top_p=0.2,
)
output = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(output)
```
### ์์
```text
**์ฌ์ฉ์:** ์ธํฐ๋ท ๋ธ๋ผ์ฐ์ ์ ๋ํด ์๋ ค์ค.
**์ธ๊ณต์ง๋ฅ:** ์ธํฐ๋ท ๋ธ๋ผ์ฐ์ ๋ ์ธํฐ๋ท์์ ์ ๋ณด๋ฅผ ๊ฒ์ํ๊ณ , ๋ค๋ฅธ ์ฌ์ฉ์์ ์ํตํ๋ ๋ฐ ์ฌ์ฉ๋๋ ์ํํธ์จ์ด์
๋๋ค.
```
```text
**์ฌ์ฉ์:** ๊ฑด๊ฐ์ ์ ์งํ๊ธฐ ์ํ ์ธ ๊ฐ์ง ํ์ ์๋ ค์ฃผ์ธ์.
**์ธ๊ณต์ง๋ฅ:** 1. ์ถฉ๋ถํ ์๋ฉด์ ์ทจํ์ธ์.
2. ๊ฑด๊ฐํ ์๋จ์ ์ญ์ทจํ์ธ์.
3. ๊ท์น์ ์ผ๋ก ์ด๋ํ์ธ์.
```
## ๋ฐ์ดํฐ์
- [maywell/koVast](https://huggingface.co/datasets/maywell/koVast)
- [KoAlpaca](https://raw.githubusercontent.com/Beomi/KoAlpaca/refs/heads/main/ko_alpaca_data.json) - ์ฝ๋, ํ๊ฐ ํฌํจ๋ ๋ฐ์ดํฐ ์ ์ธ
|