Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,13 +12,14 @@ from loguru import logger
|
|
| 12 |
from datetime import datetime
|
| 13 |
from pathlib import Path
|
| 14 |
import torch
|
| 15 |
-
from transformers import AutoProcessor, Qwen2VLForConditionalGeneration
|
|
|
|
|
|
|
| 16 |
from pdf2image import convert_from_path
|
| 17 |
import html2text
|
| 18 |
import markdown
|
| 19 |
import tempfile
|
| 20 |
|
| 21 |
-
|
| 22 |
pdf_suffixes = [".pdf"]
|
| 23 |
image_suffixes = [".png", ".jpeg", ".jpg"]
|
| 24 |
|
|
@@ -39,7 +40,7 @@ with open(header_path, 'r') as header_file:
|
|
| 39 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 40 |
MODEL_ID = "Logics-MLLM/Logics-Parsing"
|
| 41 |
processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
|
| 42 |
-
model =
|
| 43 |
MODEL_ID,
|
| 44 |
trust_remote_code=True,
|
| 45 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32
|
|
|
|
| 12 |
from datetime import datetime
|
| 13 |
from pathlib import Path
|
| 14 |
import torch
|
| 15 |
+
from transformers import AutoProcessor, Qwen2VLForConditionalGeneration, Qwen2_5_VLForConditionalGeneration
|
| 16 |
+
from transformers.image_utils import load_image
|
| 17 |
+
|
| 18 |
from pdf2image import convert_from_path
|
| 19 |
import html2text
|
| 20 |
import markdown
|
| 21 |
import tempfile
|
| 22 |
|
|
|
|
| 23 |
pdf_suffixes = [".pdf"]
|
| 24 |
image_suffixes = [".png", ".jpeg", ".jpg"]
|
| 25 |
|
|
|
|
| 40 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 41 |
MODEL_ID = "Logics-MLLM/Logics-Parsing"
|
| 42 |
processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
|
| 43 |
+
model = Qwen2_5_VLForConditionalGeneration.from_pretrained(
|
| 44 |
MODEL_ID,
|
| 45 |
trust_remote_code=True,
|
| 46 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32
|