Commit
·
2c69671
1
Parent(s):
94d0495
Update README.md
Browse files
README.md
CHANGED
|
@@ -26,10 +26,10 @@ import torch
|
|
| 26 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 27 |
|
| 28 |
|
| 29 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 30 |
|
| 31 |
model = AutoModelForCausalLM.from_pretrained(
|
| 32 |
-
"
|
| 33 |
torch_dtype=torch.bfloat16,
|
| 34 |
low_cpu_mem_usage=True,
|
| 35 |
trust_remote_code=True).to('cuda').eval()
|
|
@@ -68,10 +68,10 @@ import torch
|
|
| 68 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 69 |
|
| 70 |
|
| 71 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 72 |
|
| 73 |
model = AutoModelForCausalLM.from_pretrained(
|
| 74 |
-
"
|
| 75 |
torch_dtype=torch.bfloat16,
|
| 76 |
low_cpu_mem_usage=True,
|
| 77 |
trust_remote_code=True).to('cuda').eval()
|
|
@@ -116,11 +116,11 @@ import torch
|
|
| 116 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 117 |
from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch
|
| 118 |
|
| 119 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 120 |
|
| 121 |
with init_empty_weights():
|
| 122 |
model = AutoModelForCausalLM.from_pretrained(
|
| 123 |
-
"
|
| 124 |
torch_dtype=torch.bfloat16,
|
| 125 |
low_cpu_mem_usage=True,
|
| 126 |
trust_remote_code=True)
|
|
@@ -167,11 +167,11 @@ import torch
|
|
| 167 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 168 |
from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch
|
| 169 |
|
| 170 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 171 |
|
| 172 |
with init_empty_weights():
|
| 173 |
model = AutoModelForCausalLM.from_pretrained(
|
| 174 |
-
"
|
| 175 |
torch_dtype=torch.bfloat16,
|
| 176 |
low_cpu_mem_usage=True,
|
| 177 |
trust_remote_code=True)
|
|
@@ -226,10 +226,10 @@ import torch
|
|
| 226 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 227 |
|
| 228 |
|
| 229 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 230 |
|
| 231 |
model = AutoModelForCausalLM.from_pretrained(
|
| 232 |
-
"
|
| 233 |
load_in_4bit=True,
|
| 234 |
trust_remote_code=True,
|
| 235 |
bnb_4bit_compute_dtype=torch.float16).eval()
|
|
|
|
| 26 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 27 |
|
| 28 |
|
| 29 |
+
tokenizer = AutoTokenizer.from_pretrained("BAAI/Emu2-Chat")
|
| 30 |
|
| 31 |
model = AutoModelForCausalLM.from_pretrained(
|
| 32 |
+
"BAAI/Emu2-Chat",
|
| 33 |
torch_dtype=torch.bfloat16,
|
| 34 |
low_cpu_mem_usage=True,
|
| 35 |
trust_remote_code=True).to('cuda').eval()
|
|
|
|
| 68 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 69 |
|
| 70 |
|
| 71 |
+
tokenizer = AutoTokenizer.from_pretrained("BAAI/Emu2-Chat")
|
| 72 |
|
| 73 |
model = AutoModelForCausalLM.from_pretrained(
|
| 74 |
+
"BAAI/Emu2-Chat",
|
| 75 |
torch_dtype=torch.bfloat16,
|
| 76 |
low_cpu_mem_usage=True,
|
| 77 |
trust_remote_code=True).to('cuda').eval()
|
|
|
|
| 116 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 117 |
from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch
|
| 118 |
|
| 119 |
+
tokenizer = AutoTokenizer.from_pretrained("BAAI/Emu2-Chat")
|
| 120 |
|
| 121 |
with init_empty_weights():
|
| 122 |
model = AutoModelForCausalLM.from_pretrained(
|
| 123 |
+
"BAAI/Emu2-Chat",
|
| 124 |
torch_dtype=torch.bfloat16,
|
| 125 |
low_cpu_mem_usage=True,
|
| 126 |
trust_remote_code=True)
|
|
|
|
| 167 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 168 |
from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch
|
| 169 |
|
| 170 |
+
tokenizer = AutoTokenizer.from_pretrained("BAAI/Emu2-Chat")
|
| 171 |
|
| 172 |
with init_empty_weights():
|
| 173 |
model = AutoModelForCausalLM.from_pretrained(
|
| 174 |
+
"BAAI/Emu2-Chat",
|
| 175 |
torch_dtype=torch.bfloat16,
|
| 176 |
low_cpu_mem_usage=True,
|
| 177 |
trust_remote_code=True)
|
|
|
|
| 226 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 227 |
|
| 228 |
|
| 229 |
+
tokenizer = AutoTokenizer.from_pretrained("BAAI/Emu2-Chat")
|
| 230 |
|
| 231 |
model = AutoModelForCausalLM.from_pretrained(
|
| 232 |
+
"BAAI/Emu2-Chat",
|
| 233 |
load_in_4bit=True,
|
| 234 |
trust_remote_code=True,
|
| 235 |
bnb_4bit_compute_dtype=torch.float16).eval()
|