Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,7 +6,7 @@ import torch
|
|
| 6 |
import spaces # <-- NEW: import spaces for ZeroGPU
|
| 7 |
import gradio as gr
|
| 8 |
from PIL import Image
|
| 9 |
-
from transformers import AutoProcessor, AutoModelForVision2Seq, AutoConfig
|
| 10 |
from peft import PeftModel
|
| 11 |
import transformers, accelerate, torch, numpy as np
|
| 12 |
|
|
@@ -27,7 +27,7 @@ processor = AutoProcessor.from_pretrained(BASE_MODEL, trust_remote_code=True)
|
|
| 27 |
|
| 28 |
cfg = AutoConfig.from_pretrained(BASE_MODEL, trust_remote_code=True)
|
| 29 |
# 3. Load base model on CPU; stream shards to save RAM
|
| 30 |
-
base =
|
| 31 |
BASE_MODEL,
|
| 32 |
config=cfg,
|
| 33 |
torch_dtype=dtype,
|
|
|
|
| 6 |
import spaces # <-- NEW: import spaces for ZeroGPU
|
| 7 |
import gradio as gr
|
| 8 |
from PIL import Image
|
| 9 |
+
from transformers import AutoProcessor, AutoModelForVision2Seq, AutoConfig, AutoModelForCausalLM
|
| 10 |
from peft import PeftModel
|
| 11 |
import transformers, accelerate, torch, numpy as np
|
| 12 |
|
|
|
|
| 27 |
|
| 28 |
cfg = AutoConfig.from_pretrained(BASE_MODEL, trust_remote_code=True)
|
| 29 |
# 3. Load base model on CPU; stream shards to save RAM
|
| 30 |
+
base = AutoModelForCausalLM.from_pretrained(
|
| 31 |
BASE_MODEL,
|
| 32 |
config=cfg,
|
| 33 |
torch_dtype=dtype,
|