abbatea commited on
Commit
27e526b
·
verified ·
1 Parent(s): d9eeae0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -0
README.md CHANGED
@@ -42,6 +42,46 @@ This model is intended for use in:
42
 
43
  ---
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
  ## Citations
47
 
 
42
 
43
  ---
44
 
45
+ ## Example Usage
46
+ ```python
47
+
48
+ import torch
49
+ from transformers import AutoTokenizer, AutoModelForCausalLM
50
+ from peft import PeftModel
51
+
52
+ # Base model
53
+ basemodelname = "meta-llama/Llama-3.1-8B-Instruct"
54
+ base_model = AutoModelForCausalLM.from_pretrained(
55
+ basemodelname,
56
+ torch_dtype=torch.bfloat16,
57
+ device_map="auto"
58
+ )
59
+
60
+ # Load adapter on top
61
+ peft_model_path = "abbatea/Tutorbot-variation-DPO-Llama"
62
+ model = PeftModel.from_pretrained(base_model, peft_model_path)
63
+
64
+ # Load tokenizer
65
+ tokenizer = AutoTokenizer.from_pretrained(basemodelname)
66
+ if tokenizer.pad_token is None:
67
+ tokenizer.pad_token = tokenizer.eos_token
68
+ tokenizer.pad_token_id = tokenizer.eos_token_id
69
+
70
+
71
+
72
+ messages = [
73
+ {"role": "user", "content": "Can you help me solve 3x + 5 = 20?"}
74
+ ]
75
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False)
76
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
77
+
78
+ outputs = model.generate(**inputs, max_new_tokens=512)
79
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
+
81
+
82
+ ```
83
+ ---
84
+
85
 
86
  ## Citations
87