Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -24,7 +24,7 @@ from transformers.image_utils import load_image
|
|
| 24 |
from diffusers import StableDiffusionXLPipeline, EulerAncestralDiscreteScheduler
|
| 25 |
|
| 26 |
DESCRIPTION = """
|
| 27 |
-
# Gen Vision
|
| 28 |
"""
|
| 29 |
|
| 30 |
css = '''
|
|
@@ -47,10 +47,6 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
|
| 47 |
|
| 48 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 49 |
|
| 50 |
-
# ------------------------------
|
| 51 |
-
# Text Generation Models & TTS
|
| 52 |
-
# ------------------------------
|
| 53 |
-
|
| 54 |
# Load text-only model and tokenizer for text generation
|
| 55 |
model_id = "prithivMLmods/FastThink-0.5B-Tiny"
|
| 56 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|
@@ -329,7 +325,7 @@ demo = gr.ChatInterface(
|
|
| 329 |
description=DESCRIPTION,
|
| 330 |
css=css,
|
| 331 |
fill_height=True,
|
| 332 |
-
textbox=gr.MultimodalTextbox(label="Query Input", file_types=["image"], file_count="multiple"),
|
| 333 |
stop_btn="Stop Generation",
|
| 334 |
multimodal=True,
|
| 335 |
)
|
|
|
|
| 24 |
from diffusers import StableDiffusionXLPipeline, EulerAncestralDiscreteScheduler
|
| 25 |
|
| 26 |
DESCRIPTION = """
|
| 27 |
+
# Gen Vision ⚛️
|
| 28 |
"""
|
| 29 |
|
| 30 |
css = '''
|
|
|
|
| 47 |
|
| 48 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
# Load text-only model and tokenizer for text generation
|
| 51 |
model_id = "prithivMLmods/FastThink-0.5B-Tiny"
|
| 52 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|
|
|
| 325 |
description=DESCRIPTION,
|
| 326 |
css=css,
|
| 327 |
fill_height=True,
|
| 328 |
+
textbox=gr.MultimodalTextbox(label="Query Input", file_types=["image"], file_count="multiple", placeholder="scroll down examples to explore more art styles"),
|
| 329 |
stop_btn="Stop Generation",
|
| 330 |
multimodal=True,
|
| 331 |
)
|