DaveKevin commited on
Commit
72cde70
·
verified ·
1 Parent(s): 2e6b3d5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -73,24 +73,24 @@ from qwen_vl_utils import process_vision_info
73
 
74
  # default: Load the model on the available device(s)
75
  model = AutoModelForImageTextToText.from_pretrained(
76
- "phronetic-ai/Qwen2-VL-DAPO", torch_dtype="auto", device_map="auto"
77
  )
78
 
79
  # We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
80
  # model = AutoModelForImageTextToText.from_pretrained(
81
- # "phronetic-ai/Qwen2-VL-DAPO",
82
  # torch_dtype=torch.bfloat16,
83
  # attn_implementation="flash_attention_2",
84
  # device_map="auto",
85
  # )
86
 
87
  # default processer
88
- processor = AutoProcessor.from_pretrained("Qwen/Qwen2-VL-2B-Instruct")
89
 
90
  # The default range for the number of visual tokens per image in the model is 4-16384. You can set min_pixels and max_pixels according to your needs, such as a token count range of 256-1280, to balance speed and memory usage.
91
  # min_pixels = 256*28*28
92
  # max_pixels = 1280*28*28
93
- # processor = AutoProcessor.from_pretrained("phronetic-ai/Qwen2-VL-DAPO", min_pixels=min_pixels, max_pixels=max_pixels)
94
 
95
  messages = [
96
  {
 
73
 
74
  # default: Load the model on the available device(s)
75
  model = AutoModelForImageTextToText.from_pretrained(
76
+ "phronetic-ai/RZN-V", torch_dtype="auto", device_map="auto"
77
  )
78
 
79
  # We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
80
  # model = AutoModelForImageTextToText.from_pretrained(
81
+ # "phronetic-ai/RZN-V",
82
  # torch_dtype=torch.bfloat16,
83
  # attn_implementation="flash_attention_2",
84
  # device_map="auto",
85
  # )
86
 
87
  # default processer
88
+ processor = AutoProcessor.from_pretrained("phronetic-ai/RZN-V")
89
 
90
  # The default range for the number of visual tokens per image in the model is 4-16384. You can set min_pixels and max_pixels according to your needs, such as a token count range of 256-1280, to balance speed and memory usage.
91
  # min_pixels = 256*28*28
92
  # max_pixels = 1280*28*28
93
+ # processor = AutoProcessor.from_pretrained("phronetic-ai/RZN-V", min_pixels=min_pixels, max_pixels=max_pixels)
94
 
95
  messages = [
96
  {