Update README.md
Browse files
README.md
CHANGED
|
@@ -45,7 +45,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
| 45 |
class TextGenerationAssistant:
|
| 46 |
def __init__(self, model_id:str):
|
| 47 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 48 |
-
self.model = AutoModelForCausalLM.from_pretrained(model_id, device_map='auto'
|
| 49 |
self.pipe = pipeline("text-generation",
|
| 50 |
model=self.model,
|
| 51 |
tokenizer=self.tokenizer,
|
|
@@ -56,10 +56,10 @@ class TextGenerationAssistant:
|
|
| 56 |
)
|
| 57 |
|
| 58 |
self.sampling_params = dict(do_sample=True, temperature=0.5, top_k=50, top_p=0.9)
|
| 59 |
-
self.
|
| 60 |
|
| 61 |
def format_prompt(self, user_input):
|
| 62 |
-
return "[INST] " + self.
|
| 63 |
|
| 64 |
def generate_response(self, user_query):
|
| 65 |
prompt = self.format_prompt(user_query)
|
|
|
|
| 45 |
class TextGenerationAssistant:
|
| 46 |
def __init__(self, model_id:str):
|
| 47 |
self.tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 48 |
+
self.model = AutoModelForCausalLM.from_pretrained(model_id, device_map='auto')
|
| 49 |
self.pipe = pipeline("text-generation",
|
| 50 |
model=self.model,
|
| 51 |
tokenizer=self.tokenizer,
|
|
|
|
| 56 |
)
|
| 57 |
|
| 58 |
self.sampling_params = dict(do_sample=True, temperature=0.5, top_k=50, top_p=0.9)
|
| 59 |
+
self.system_prompt = "Sen yardımcı bir asistansın. Sana verilen talimat ve girdilere en uygun cevapları üreteceksin. \n\n\n"
|
| 60 |
|
| 61 |
def format_prompt(self, user_input):
|
| 62 |
+
return "[INST] " + self.system_prompt + user_input + " [/INST]"
|
| 63 |
|
| 64 |
def generate_response(self, user_query):
|
| 65 |
prompt = self.format_prompt(user_query)
|