Update README.md
Browse files
README.md
CHANGED
|
@@ -70,7 +70,7 @@ class TextGenerationAssistant:
|
|
| 70 |
assistant = TextGenerationAssistant(model_id="Commencis/Commencis-LLM")
|
| 71 |
|
| 72 |
# Enter your query here.
|
| 73 |
-
user_query = "Faiz
|
| 74 |
response = assistant.generate_response(user_query)
|
| 75 |
print(response)
|
| 76 |
|
|
@@ -84,7 +84,7 @@ import transformers
|
|
| 84 |
import torch
|
| 85 |
|
| 86 |
model = "Commencis/Commencis-LLM"
|
| 87 |
-
messages = [{"role": "user", "content": "Faiz
|
| 88 |
|
| 89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
| 90 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
|
|
|
| 70 |
assistant = TextGenerationAssistant(model_id="Commencis/Commencis-LLM")
|
| 71 |
|
| 72 |
# Enter your query here.
|
| 73 |
+
user_query = "Faiz oranı yükseldiğinde kredi maliyetim nasıl etkilenir?"
|
| 74 |
response = assistant.generate_response(user_query)
|
| 75 |
print(response)
|
| 76 |
|
|
|
|
| 84 |
import torch
|
| 85 |
|
| 86 |
model = "Commencis/Commencis-LLM"
|
| 87 |
+
messages = [{"role": "user", "content": "Faiz oranı yükseldiğinde kredi maliyetim nasıl etkilenir?"}]
|
| 88 |
|
| 89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
| 90 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|