zhibinlan commited on
Commit
391930a
·
1 Parent(s): caab229
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -57,13 +57,13 @@ from qwen_vl_utils import process_vision_info
57
  import torch
58
 
59
  model = Qwen2VLForConditionalGeneration.from_pretrained(
60
- "zhibinlan/UME-R1-2B",
61
  torch_dtype=torch.bfloat16,
62
  attn_implementation="flash_attention_2",
63
  device_map="cuda:0",
64
  )
65
 
66
- processor = AutoProcessor.from_pretrained("zhibinlan/UME-R1-2B")
67
 
68
  prompt = '''Represent the above input text, images, videos, or any combination of the three as embeddings.
69
  First output the thinking process in <think> </think> tags and then summarize the entire input in a word or sentence.
@@ -148,7 +148,7 @@ from transformers import Qwen2VLForConditionalGeneration,AutoProcessor
148
  from qwen_vl_utils import process_vision_info
149
  import torch
150
 
151
- pretrained_path = "release/UME-R1-2B"
152
 
153
  # We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
154
  model = Qwen2VLForConditionalGeneration.from_pretrained(
 
57
  import torch
58
 
59
  model = Qwen2VLForConditionalGeneration.from_pretrained(
60
+ "zhibinlan/UME-R1-7B",
61
  torch_dtype=torch.bfloat16,
62
  attn_implementation="flash_attention_2",
63
  device_map="cuda:0",
64
  )
65
 
66
+ processor = AutoProcessor.from_pretrained("zhibinlan/UME-R1-7B")
67
 
68
  prompt = '''Represent the above input text, images, videos, or any combination of the three as embeddings.
69
  First output the thinking process in <think> </think> tags and then summarize the entire input in a word or sentence.
 
148
  from qwen_vl_utils import process_vision_info
149
  import torch
150
 
151
+ pretrained_path = "zhibinlan/UME-R1-7B"
152
 
153
  # We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
154
  model = Qwen2VLForConditionalGeneration.from_pretrained(