Update README.md
Browse files
README.md
CHANGED
|
@@ -37,7 +37,7 @@ import transformers
|
|
| 37 |
from transformers import PeftModel
|
| 38 |
|
| 39 |
model_name = "google/flan-t5-large"
|
| 40 |
-
peft_model_id = "
|
| 41 |
tokenizer = transformers.AutoTokenizer.from_pretrained(model_name)
|
| 42 |
base_model = transformers.AutoModelForSeq2SeqLM.from_pretrained(model_name)
|
| 43 |
peft_model = PeftModel.from_pretrained(base_model, peft_model_id)
|
|
@@ -45,9 +45,3 @@ peft_model = PeftModel.from_pretrained(base_model, peft_model_id)
|
|
| 45 |
inputs = tokenizer("""[INSERT INSTRUCTION HERE]""", return_tensors="pt")
|
| 46 |
outputs = peft_model.generate(**inputs, max_length=300, do_sample=True)
|
| 47 |
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
---
|
| 52 |
-
license: apache-2.0
|
| 53 |
-
---
|
|
|
|
| 37 |
from transformers import PeftModel
|
| 38 |
|
| 39 |
model_name = "google/flan-t5-large"
|
| 40 |
+
peft_model_id = "Leadmatic/tinyChat"
|
| 41 |
tokenizer = transformers.AutoTokenizer.from_pretrained(model_name)
|
| 42 |
base_model = transformers.AutoModelForSeq2SeqLM.from_pretrained(model_name)
|
| 43 |
peft_model = PeftModel.from_pretrained(base_model, peft_model_id)
|
|
|
|
| 45 |
inputs = tokenizer("""[INSERT INSTRUCTION HERE]""", return_tensors="pt")
|
| 46 |
outputs = peft_model.generate(**inputs, max_length=300, do_sample=True)
|
| 47 |
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|