Update README.md
Browse files
README.md
CHANGED
|
@@ -26,7 +26,7 @@ import torch
|
|
| 26 |
from transformers import AutoModelForCausalLM, AutoProcessor
|
| 27 |
from open_flamingo.eval.models.cruise_model import EvalModel
|
| 28 |
|
| 29 |
-
processor = AutoProcessor.from_pretrained("
|
| 30 |
|
| 31 |
prompts = [
|
| 32 |
{
|
|
@@ -40,7 +40,7 @@ prompts = [
|
|
| 40 |
inputs = processor(prompts)
|
| 41 |
# use bf16 and gpu 0
|
| 42 |
model = AutoModelForCausalLM.from_pretrained(
|
| 43 |
-
"
|
| 44 |
local_files_only=True,
|
| 45 |
torch_dtype=torch.bfloat16,
|
| 46 |
trust_remote_code=True,
|
|
|
|
| 26 |
from transformers import AutoModelForCausalLM, AutoProcessor
|
| 27 |
from open_flamingo.eval.models.cruise_model import EvalModel
|
| 28 |
|
| 29 |
+
processor = AutoProcessor.from_pretrained("Infi-MM/infimm-hd", trust_remote_code=True)
|
| 30 |
|
| 31 |
prompts = [
|
| 32 |
{
|
|
|
|
| 40 |
inputs = processor(prompts)
|
| 41 |
# use bf16 and gpu 0
|
| 42 |
model = AutoModelForCausalLM.from_pretrained(
|
| 43 |
+
"Infi-MM/infimm-hd",
|
| 44 |
local_files_only=True,
|
| 45 |
torch_dtype=torch.bfloat16,
|
| 46 |
trust_remote_code=True,
|