DevSrijit commited on
Commit
ed3274d
·
verified ·
1 Parent(s): c40061f

Create model.py

Browse files
Files changed (1) hide show
  1. app/model.py +21 -0
app/model.py ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer
2
+ import torch
3
+
4
+ class ModelManager:
5
+ def __init__(self, model_name="BanglaLLM/BanglaLLama-3.2-3b-unlop-culturax-base-v0.0.3"):
6
+ self.device = "cuda" if torch.cuda.is_available() else "cpu"
7
+ self.tokenizer = AutoTokenizer.from_pretrained(model_name)
8
+ self.model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16 if self.device == "cuda" else torch.float32)
9
+ self.model = self.model.to(self.device)
10
+
11
+ def generate_response(self, prompt, max_length, temperature, top_k, top_p):
12
+ inputs = self.tokenizer(prompt, return_tensors="pt").to(self.device)
13
+ outputs = self.model.generate(
14
+ inputs["input_ids"],
15
+ max_length=max_length,
16
+ temperature=temperature,
17
+ top_k=top_k,
18
+ top_p=top_p,
19
+ do_sample=True
20
+ )
21
+ return self.tokenizer.decode(outputs[0], skip_special_tokens=True)