Prompt Tempalte

It follows Alpaca format.

### 질문: {instruction}
### λ‹΅λ³€: {output}

Implementation Code

import troch
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.fron_pretrained("Ja3ck/Mistral-instruct-IPO-Y24-v1", return_dict=True, torch_dtype=torch.bfloat16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("Ja3ck/Mistral-instruct-IPO-Y24-v1", use_fast=True)
tokenizer.pad_token = tokenizer.unk_token
tokenizer.pad_token_id = tokenizer.unk_token_id
tokenizer.padding_side = "left"

def gen(x):
  x_ = f"### 질문: {x.strip()} ### λ‹΅λ³€: "
  inputs = tokenizer(x_, return_tensor='pt')
  input_ids = inputs['input_ids'].cuda()
  generation_output = model.generate(
      pad_token_id = tokenizer.pad_token_id,
      temperature=0.1,
      top_p=1,
      top_k=50,
      num_beams=1,
      repetition_penalty=1.13,
      do_sample=True,
    ),
    return_dict_in_generate=True,
    output_scores=True,
    max_new_tokens=1024
  )
  for seq in generation_output.sequences:
    output = tokenizer.decode(seq)
    print(output.split("### λ‹΅λ³€: ")[1].strip())

gen("μ•ˆλ…•ν•˜μ„Έμš”?")
Downloads last month
4
Safetensors
Model size
7B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Ja-ck/Mistral-instruct-IPO-Y24-v1

Quantizations
1 model

Spaces using Ja-ck/Mistral-instruct-IPO-Y24-v1 7