Update README.md
Browse files
README.md
CHANGED
|
@@ -93,7 +93,7 @@ from transformers import AutoConfig, AutoModelForCausalLM
|
|
| 93 |
repo_name = "nvidia/Nemotron-Flash-3B"
|
| 94 |
|
| 95 |
config = AutoConfig.from_pretrained(repo_name, trust_remote_code=True)
|
| 96 |
-
setattr(config, "
|
| 97 |
model = AutoModelForCausalLM.from_pretrained(repo_name, config=config, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
| 98 |
```
|
| 99 |
|
|
|
|
| 93 |
repo_name = "nvidia/Nemotron-Flash-3B"
|
| 94 |
|
| 95 |
config = AutoConfig.from_pretrained(repo_name, trust_remote_code=True)
|
| 96 |
+
setattr(config, "attn_implementation_new", "flash_attention_2")
|
| 97 |
model = AutoModelForCausalLM.from_pretrained(repo_name, config=config, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
| 98 |
```
|
| 99 |
|