Korean LLM
Collection
Open source LLMs for korean โข 10 items โข Updated โข 1
This model is a Supervised fine-tuned version of x2bee/POLAR-14B-v0.2 with DeepSpeed and trl for korean.
from transformers import TextStreamer, pipeline, AutoTokenizer, AutoModelForCausalLM
model_id = 'spow12/POLAR-14B_4.3_very_big_sft'
tokenizer = AutoTokenizer.from_pretrained(model_id)
# %%
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
attn_implementation="flash_attention_2",
device_map='auto',
)
model.eval()
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, device_map='auto')
streamer = TextStreamer(tokenizer)
generation_configs = dict(
max_new_tokens=2048,
num_return_sequences=1,
temperature=0.1,
# early_stopping=True,
repetition_penalty=1.2,
num_beams=1,
do_sample=True,
top_k=20,
top_p=0.9,
eos_token_id=tokenizer.eos_token_id,
pad_token_id=tokenizer.eos_token_id,
streamer=streamer
)
sys_message = """๋น์ ์ ์น์ ํ ์ฑ๋ด์ผ๋ก์ ์๋๋ฐฉ์ ์์ฒญ์ ์ต๋ํ ์์ธํ๊ณ ์น์ ํ๊ฒ ๋ตํด์ผํฉ๋๋ค.
์ฌ์ฉ์๊ฐ ์ ๊ณตํ๋ ์ ๋ณด๋ฅผ ์ธ์ฌํ๊ฒ ๋ถ์ํ์ฌ ์ฌ์ฉ์์ ์๋๋ฅผ ์ ์ํ๊ฒ ํ์
ํ๊ณ ๊ทธ์ ๋ฐ๋ผ ๋ต๋ณ์ ์์ฑํด์ผํฉ๋๋ค.
ํญ์ ๋งค์ฐ ์์ฐ์ค๋ฌ์ด ํ๊ตญ์ด๋ก ์๋ตํ์ธ์."""
message = [
{
'role': "system",
'content': sys_message
},
{
'role': 'user',
'content': "ํ์ฌ์ ๊ฒฝ์ ์ํฉ์ ๋ํด ์ด๋ป๊ฒ ์๊ฐํด?."
}
]
conversation = pipe(message, **generation_configs)
conversation[-1]
This model is licensed under the cc-by-nc-4.0. which allows others to share and adapt the model for non-commercial purposes.
Here is Original Readme.md