VaibhavHD commited on
Commit
fc2e2dc
·
verified ·
1 Parent(s): f2903c8

Update inference.py

Browse files
Files changed (1) hide show
  1. inference.py +14 -14
inference.py CHANGED
@@ -1,14 +1,14 @@
1
- from transformers import AutoTokenizer, AutoModelForCausalLM
2
- from peft import PeftModel
3
-
4
- BASE_MODEL = "deepseek-ai/deepseek-coder-1.3b-base"
5
- LORA_REPO = "your-username/deepseek-lora-monthly"
6
-
7
- tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
8
- base = AutoModelForCausalLM.from_pretrained(BASE_MODEL, trust_remote_code=True)
9
- model = PeftModel.from_pretrained(base, LORA_REPO)
10
-
11
- def generate_response(prompt:str)->str:
12
- inputs = tokenizer(prompt, return_tensors="pt")
13
- out = model.generate(**inputs, max_new_tokens=200)
14
- return tokenizer.decode(out[0], skip_special_tokens=True)
 
1
+ from transformers import AutoTokenizer, AutoModelForCausalLM
2
+ from peft import PeftModel
3
+
4
+ BASE_MODEL = "deepseek-ai/deepseek-coder-1.3b-base"
5
+ LORA_REPO = "https://huggingface.co/VaibhavHD/deepseek-lora-monthly"
6
+
7
+ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
8
+ base = AutoModelForCausalLM.from_pretrained(BASE_MODEL, trust_remote_code=True)
9
+ model = PeftModel.from_pretrained(base, LORA_REPO)
10
+
11
+ def generate_response(prompt:str)->str:
12
+ inputs = tokenizer(prompt, return_tensors="pt")
13
+ out = model.generate(**inputs, max_new_tokens=200)
14
+ return tokenizer.decode(out[0], skip_special_tokens=True)