Update README.md
Browse files
README.md
CHANGED
|
@@ -105,20 +105,20 @@ from conv_for_infer import generate_conv
|
|
| 105 |
|
| 106 |
# Load the model
|
| 107 |
model = Qwen3VLForConditionalGeneration.from_pretrained(
|
| 108 |
-
"FireRedTeam/FireRed-OCR
|
| 109 |
torch_dtype=torch.bfloat16,
|
| 110 |
device_map="auto",
|
| 111 |
)
|
| 112 |
|
| 113 |
# We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
|
| 114 |
# model = Qwen3VLForConditionalGeneration.from_pretrained(
|
| 115 |
-
# "
|
| 116 |
# dtype=torch.bfloat16,
|
| 117 |
# attn_implementation="flash_attention_2",
|
| 118 |
# device_map="auto",
|
| 119 |
# )
|
| 120 |
|
| 121 |
-
processor = AutoProcessor.from_pretrained("FireRedTeam/FireRed-OCR
|
| 122 |
|
| 123 |
# Prepare Input
|
| 124 |
image_path = "./examples/complex_table.png"
|
|
|
|
| 105 |
|
| 106 |
# Load the model
|
| 107 |
model = Qwen3VLForConditionalGeneration.from_pretrained(
|
| 108 |
+
"FireRedTeam/FireRed-OCR",
|
| 109 |
torch_dtype=torch.bfloat16,
|
| 110 |
device_map="auto",
|
| 111 |
)
|
| 112 |
|
| 113 |
# We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
|
| 114 |
# model = Qwen3VLForConditionalGeneration.from_pretrained(
|
| 115 |
+
# "FireRedTeam/FireRed-OCR,
|
| 116 |
# dtype=torch.bfloat16,
|
| 117 |
# attn_implementation="flash_attention_2",
|
| 118 |
# device_map="auto",
|
| 119 |
# )
|
| 120 |
|
| 121 |
+
processor = AutoProcessor.from_pretrained("FireRedTeam/FireRed-OCR")
|
| 122 |
|
| 123 |
# Prepare Input
|
| 124 |
image_path = "./examples/complex_table.png"
|