YongganFu commited on
Commit
e623d13
·
verified ·
1 Parent(s): 972cbc6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -93,7 +93,7 @@ from transformers import AutoConfig, AutoModelForCausalLM
93
  repo_name = "nvidia/Nemotron-Flash-3B"
94
 
95
  config = AutoConfig.from_pretrained(repo_name, trust_remote_code=True)
96
- setattr(config, "attention_implementation_new", "flash_attention_2")
97
  model = AutoModelForCausalLM.from_pretrained(repo_name, config=config, torch_dtype=torch.bfloat16, trust_remote_code=True)
98
  ```
99
 
 
93
  repo_name = "nvidia/Nemotron-Flash-3B"
94
 
95
  config = AutoConfig.from_pretrained(repo_name, trust_remote_code=True)
96
+ setattr(config, "attn_implementation_new", "flash_attention_2")
97
  model = AutoModelForCausalLM.from_pretrained(repo_name, config=config, torch_dtype=torch.bfloat16, trust_remote_code=True)
98
  ```
99