Update README.md
Browse files
README.md
CHANGED
|
@@ -53,14 +53,14 @@ double_quant_config = BitsAndBytesConfig(
|
|
| 53 |
)
|
| 54 |
|
| 55 |
model = AutoModelForCausalLM.from_pretrained(
|
| 56 |
-
"Severian/Jamba-
|
| 57 |
device_map="auto",
|
| 58 |
trust_remote_code=True,
|
| 59 |
torch_dtype=torch.bfloat16,
|
| 60 |
attn_implementation="flash_attention_2",
|
| 61 |
quantization_config=double_quant_config,
|
| 62 |
)
|
| 63 |
-
tokenizer = AutoTokenizer.from_pretrained("Severian/Jamba-
|
| 64 |
|
| 65 |
input_text = """<|im_start|>system
|
| 66 |
You are a helpful AI assistant.
|
|
|
|
| 53 |
)
|
| 54 |
|
| 55 |
model = AutoModelForCausalLM.from_pretrained(
|
| 56 |
+
"Severian/Jamba-Open-Hermes",
|
| 57 |
device_map="auto",
|
| 58 |
trust_remote_code=True,
|
| 59 |
torch_dtype=torch.bfloat16,
|
| 60 |
attn_implementation="flash_attention_2",
|
| 61 |
quantization_config=double_quant_config,
|
| 62 |
)
|
| 63 |
+
tokenizer = AutoTokenizer.from_pretrained("Severian/Jamba-Open-Hermes")
|
| 64 |
|
| 65 |
input_text = """<|im_start|>system
|
| 66 |
You are a helpful AI assistant.
|