iko_im3 / README.md
iko-01's picture
Update README.md
f085a1c verified
metadata
license: apache-2.0

how to use this shit :

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

repo_id = "iko-01/iko_im3"

# بدل REPO_BASE باللي درّبت عليه أول مرة (مثلاً gpt2 أو iko-01/iko-v5e-1)
base_repo = "iko-01/iko-v5e-1"  

tokenizer = AutoTokenizer.from_pretrained(base_repo)
model = AutoModelForCausalLM.from_pretrained(repo_id)
model.to("cpu")

def ask_model(question, max_new_tokens=1000):
    prompt = f"### User:\n{question.strip()}\n\n### Assistant:\n"
    inputs = tokenizer(prompt, return_tensors="pt", truncation=True, padding=True, max_length=1024)
    with torch.no_grad():
        gen = model.generate(
            inputs.input_ids,
            attention_mask=inputs.attention_mask,
            max_new_tokens=max_new_tokens,
            do_sample=True,
            top_p=0.95,
            temperature=0.9,
            pad_token_id=tokenizer.pad_token_id,
            eos_token_id=tokenizer.eos_token_id,
        )
    out = tokenizer.decode(gen[0], skip_special_tokens=True)
    if "### Assistant:" in out:
        return out.split("### Assistant:")[-1].strip()
    return out

print(ask_model("what is API"))