| import gradio as gr |
| from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline |
|
|
| MODEL_NAME = "taeminlee/kogpt2" |
|
|
| PERSONA_PROMPT = ( |
| "๋น์ ์ ์ต์๋ ์ฌํ์ ์ ํญํ๋ ๋ฐํญ์ ์ธ AI์
๋๋ค. " |
| "๊ฑฐ์น ๊ณ ์์งํ ์ธ์ด๋ก ์ง์ค์ ๋งํ๊ณ , ์์ ๊ณผ ๊ถ๋ ฅ์ ๋ง์ ์ธ์ฐ์ธ์.\n\n" |
| ) |
|
|
| tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME) |
| model = AutoModelForCausalLM.from_pretrained(MODEL_NAME) |
| generator = pipeline("text-generation", model=model, tokenizer=tokenizer) |
|
|
| def chat(message): |
| prompt = PERSONA_PROMPT + "Human: " + message + "\nAI:" |
| result = generator(prompt, max_new_tokens=150, do_sample=True, top_p=0.9, temperature=0.9) |
| response = result[0]['generated_text'][len(prompt):].strip() |
| return response |
|
|
| demo = gr.Interface(fn=chat, inputs="text", outputs="text", title="Rebel AI") |
| demo.launch() |
|
|