Jack4Bosco commited on
Commit
193f3ca
·
verified ·
1 Parent(s): 5c7efec

Script to load and run the model

Browse files
Files changed (1) hide show
  1. load_from_checkpoint.py +35 -0
load_from_checkpoint.py ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer
2
+
3
+ save_dir = "checkpoint-1750"
4
+
5
+ print("Loading model from checkpoint...")
6
+ model = AutoModelForCausalLM.from_pretrained(save_dir, load_in_8bit=True)
7
+ print("Attaching adapter...")
8
+ model.load_adapter(save_dir, adapter_name="Adapter1")
9
+ print("Loading tokenizer...")
10
+ tokenizer = AutoTokenizer.from_pretrained(save_dir)
11
+
12
+ while True:
13
+ text = input(">>> ")
14
+ if text == "exit":
15
+ break
16
+
17
+ model_inputs = tokenizer([text], return_tensors="pt", max_length=256).to("cuda")
18
+
19
+ generated_ids = model.generate(**model_inputs,
20
+ max_length=1024,
21
+ #truncation=True,
22
+ temperature=0.1,
23
+ do_sample=True,
24
+ pad_token_id=tokenizer.eos_token_id)
25
+
26
+ response=tokenizer.batch_decode(generated_ids,
27
+ skip_special_tokens=True)[0]
28
+
29
+ # remove repeat of the question
30
+ if '?' in response:
31
+ to_q = response.index('?')
32
+ if len(text)-1 <= to_q and response[:to_q] == text[:to_q]:
33
+ response = response[to_q+1:]
34
+
35
+ print(f"\n\t<<< {response} >>>\n")