YimingZeng commited on
Commit
aa3901d
·
verified ·
1 Parent(s): 2f6484e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -2
README.md CHANGED
@@ -1,5 +1,46 @@
1
  ---
2
  license: bigscience-openrail-m
3
  base_model:
4
- - Qwen/Qwen2.5-3B-Instruct
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: bigscience-openrail-m
3
  base_model:
4
+ - meta-llama/Llama-3.2-3B
5
+ - Qwen/Qwen2.5-3B-Instruct
6
+ ---
7
+
8
+
9
+ The following example illustrates how to load the FineEdit-XL adapter for inference.
10
+
11
+ ```python
12
+ from transformers import AutoModelForCausalLM, AutoTokenizer
13
+ from peft import PeftModel
14
+ import torch
15
+
16
+ ## Base Model
17
+ base_model = "meta-llama/Llama-3.2-3B"
18
+
19
+ ## Lora
20
+ adapter_model = "YimingZeng/FineEdit_Model"
21
+
22
+ subfolder = "FineEdit-XL"
23
+
24
+ ## Load tokenizer and base model
25
+ tokenizer = AutoTokenizer.from_pretrained(base_model)
26
+ base = AutoModelForCausalLM.from_pretrained(
27
+ base_model,
28
+ torch_dtype=torch.bfloat16,
29
+ device_map="auto"
30
+ )
31
+
32
+ ## Load LoRA adapter
33
+ model = PeftModel.from_pretrained(base, adapter_model, subfolder=subfolder)
34
+
35
+ ## Test
36
+ prompt = """Edit Request: Please change 'Captain American' to 'Iron Man'.
37
+ Original Content: 'Captain American' is a superhero.
38
+ Edited Content:
39
+ """
40
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
41
+
42
+ with torch.no_grad():
43
+ outputs = model.generate(**inputs, max_new_tokens=100)
44
+
45
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
46
+ ```