Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ from loguru import logger
|
|
| 14 |
from PIL import Image
|
| 15 |
from transformers import AutoProcessor, AutoModelForImageTextToText, TextIteratorStreamer
|
| 16 |
|
| 17 |
-
model_id = os.getenv("MODEL_ID", "google/medgemma-
|
| 18 |
processor = AutoProcessor.from_pretrained(model_id)
|
| 19 |
model = AutoModelForImageTextToText.from_pretrained(
|
| 20 |
model_id, device_map="auto", torch_dtype=torch.bfloat16
|
|
@@ -218,7 +218,7 @@ demo = gr.ChatInterface(
|
|
| 218 |
gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
|
| 219 |
],
|
| 220 |
stop_btn=False,
|
| 221 |
-
title="MedGemma
|
| 222 |
description=DESCRIPTION,
|
| 223 |
run_examples_on_click=False,
|
| 224 |
cache_examples=False,
|
|
|
|
| 14 |
from PIL import Image
|
| 15 |
from transformers import AutoProcessor, AutoModelForImageTextToText, TextIteratorStreamer
|
| 16 |
|
| 17 |
+
model_id = os.getenv("MODEL_ID", "google/medgemma-27b-it")
|
| 18 |
processor = AutoProcessor.from_pretrained(model_id)
|
| 19 |
model = AutoModelForImageTextToText.from_pretrained(
|
| 20 |
model_id, device_map="auto", torch_dtype=torch.bfloat16
|
|
|
|
| 218 |
gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
|
| 219 |
],
|
| 220 |
stop_btn=False,
|
| 221 |
+
title="MedGemma 27B IT",
|
| 222 |
description=DESCRIPTION,
|
| 223 |
run_examples_on_click=False,
|
| 224 |
cache_examples=False,
|