hssling commited on
Commit
c238547
·
1 Parent(s): 27ed2ac

Fix transformers import for Qwen2-VL

Browse files
Files changed (2) hide show
  1. app.py +2 -2
  2. train_multimodal.py +2 -2
app.py CHANGED
@@ -1,6 +1,6 @@
1
  import gradio as gr
2
  import torch
3
- from transformers import AutoProcessor, AutoModelForVision2Seq
4
  from PIL import Image
5
 
6
  # 1. HuggingFace Space Deployment Settings
@@ -16,7 +16,7 @@ print(f"Loading {MODEL_ID}...")
16
 
17
  device = "cuda" if torch.cuda.is_available() else "cpu"
18
  processor = AutoProcessor.from_pretrained(MODEL_ID)
19
- model = AutoModelForVision2Seq.from_pretrained(
20
  MODEL_ID,
21
  torch_dtype=torch.float16 if device == "cuda" else torch.float32,
22
  device_map="auto"
 
1
  import gradio as gr
2
  import torch
3
+ from transformers import AutoProcessor, Qwen2VLForConditionalGeneration
4
  from PIL import Image
5
 
6
  # 1. HuggingFace Space Deployment Settings
 
16
 
17
  device = "cuda" if torch.cuda.is_available() else "cpu"
18
  processor = AutoProcessor.from_pretrained(MODEL_ID)
19
+ model = Qwen2VLForConditionalGeneration.from_pretrained(
20
  MODEL_ID,
21
  torch_dtype=torch.float16 if device == "cuda" else torch.float32,
22
  device_map="auto"
train_multimodal.py CHANGED
@@ -1,5 +1,5 @@
1
  import torch
2
- from transformers import AutoProcessor, AutoModelForVision2Seq, TrainingArguments
3
  from peft import LoraConfig, get_peft_model
4
  from datasets import load_dataset
5
  from trl import SFTTrainer
@@ -28,7 +28,7 @@ def main():
28
  # Load processor and model with memory-efficient 4-bit quantization
29
  processor = AutoProcessor.from_pretrained(MODEL_ID)
30
 
31
- model = AutoModelForVision2Seq.from_pretrained(
32
  MODEL_ID,
33
  device_map="auto",
34
  torch_dtype=torch.float16,
 
1
  import torch
2
+ from transformers import AutoProcessor, Qwen2VLForConditionalGeneration, TrainingArguments
3
  from peft import LoraConfig, get_peft_model
4
  from datasets import load_dataset
5
  from trl import SFTTrainer
 
28
  # Load processor and model with memory-efficient 4-bit quantization
29
  processor = AutoProcessor.from_pretrained(MODEL_ID)
30
 
31
+ model = Qwen2VLForConditionalGeneration.from_pretrained(
32
  MODEL_ID,
33
  device_map="auto",
34
  torch_dtype=torch.float16,