Update README.md
Browse files
README.md
CHANGED
|
@@ -4,9 +4,6 @@ tags:
|
|
| 4 |
- mergekit
|
| 5 |
- gemma-2b
|
| 6 |
pipeline_tag: text-generation
|
| 7 |
-
language:
|
| 8 |
-
- ar
|
| 9 |
-
- en
|
| 10 |
---
|
| 11 |
|
| 12 |
# Merge-Gemma-2b-Models
|
|
@@ -42,11 +39,10 @@ import transformers
|
|
| 42 |
import torch
|
| 43 |
|
| 44 |
model = "Ruqiya/Merge-Gemma-2b-Models"
|
| 45 |
-
|
| 46 |
-
messagesEn = [{"role": "user", "content": "What is artificial intelligence?"}]
|
| 47 |
|
| 48 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
| 49 |
-
prompt = tokenizer.apply_chat_template(
|
| 50 |
pipeline = transformers.pipeline(
|
| 51 |
"text-generation",
|
| 52 |
model=model,
|
|
|
|
| 4 |
- mergekit
|
| 5 |
- gemma-2b
|
| 6 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
# Merge-Gemma-2b-Models
|
|
|
|
| 39 |
import torch
|
| 40 |
|
| 41 |
model = "Ruqiya/Merge-Gemma-2b-Models"
|
| 42 |
+
messages = [{"role": "user", "content": "What is artificial intelligence?"}]
|
|
|
|
| 43 |
|
| 44 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
| 45 |
+
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 46 |
pipeline = transformers.pipeline(
|
| 47 |
"text-generation",
|
| 48 |
model=model,
|