zhibinlan
commited on
Commit
·
391930a
1
Parent(s):
caab229
update
Browse files
README.md
CHANGED
|
@@ -57,13 +57,13 @@ from qwen_vl_utils import process_vision_info
|
|
| 57 |
import torch
|
| 58 |
|
| 59 |
model = Qwen2VLForConditionalGeneration.from_pretrained(
|
| 60 |
-
"zhibinlan/UME-R1-
|
| 61 |
torch_dtype=torch.bfloat16,
|
| 62 |
attn_implementation="flash_attention_2",
|
| 63 |
device_map="cuda:0",
|
| 64 |
)
|
| 65 |
|
| 66 |
-
processor = AutoProcessor.from_pretrained("zhibinlan/UME-R1-
|
| 67 |
|
| 68 |
prompt = '''Represent the above input text, images, videos, or any combination of the three as embeddings.
|
| 69 |
First output the thinking process in <think> </think> tags and then summarize the entire input in a word or sentence.
|
|
@@ -148,7 +148,7 @@ from transformers import Qwen2VLForConditionalGeneration,AutoProcessor
|
|
| 148 |
from qwen_vl_utils import process_vision_info
|
| 149 |
import torch
|
| 150 |
|
| 151 |
-
pretrained_path = "
|
| 152 |
|
| 153 |
# We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
|
| 154 |
model = Qwen2VLForConditionalGeneration.from_pretrained(
|
|
|
|
| 57 |
import torch
|
| 58 |
|
| 59 |
model = Qwen2VLForConditionalGeneration.from_pretrained(
|
| 60 |
+
"zhibinlan/UME-R1-7B",
|
| 61 |
torch_dtype=torch.bfloat16,
|
| 62 |
attn_implementation="flash_attention_2",
|
| 63 |
device_map="cuda:0",
|
| 64 |
)
|
| 65 |
|
| 66 |
+
processor = AutoProcessor.from_pretrained("zhibinlan/UME-R1-7B")
|
| 67 |
|
| 68 |
prompt = '''Represent the above input text, images, videos, or any combination of the three as embeddings.
|
| 69 |
First output the thinking process in <think> </think> tags and then summarize the entire input in a word or sentence.
|
|
|
|
| 148 |
from qwen_vl_utils import process_vision_info
|
| 149 |
import torch
|
| 150 |
|
| 151 |
+
pretrained_path = "zhibinlan/UME-R1-7B"
|
| 152 |
|
| 153 |
# We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
|
| 154 |
model = Qwen2VLForConditionalGeneration.from_pretrained(
|