Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -15,12 +15,12 @@ repo = "stabilityai/stable-diffusion-3.5-large"
|
|
| 15 |
t2i = StableDiffusion3Pipeline.from_pretrained(repo, torch_dtype=torch.bfloat16, token=os.environ["TOKEN"]).to(device)
|
| 16 |
|
| 17 |
model = AutoModelForCausalLM.from_pretrained(
|
| 18 |
-
"
|
| 19 |
device_map="cuda",
|
| 20 |
torch_dtype=torch.bfloat16,
|
| 21 |
trust_remote_code=True,
|
| 22 |
)
|
| 23 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
| 24 |
upsampler = pipeline(
|
| 25 |
"text-generation",
|
| 26 |
model=model,
|
|
|
|
| 15 |
t2i = StableDiffusion3Pipeline.from_pretrained(repo, torch_dtype=torch.bfloat16, token=os.environ["TOKEN"]).to(device)
|
| 16 |
|
| 17 |
model = AutoModelForCausalLM.from_pretrained(
|
| 18 |
+
"google/gemma-2-2b-jpn-it",
|
| 19 |
device_map="cuda",
|
| 20 |
torch_dtype=torch.bfloat16,
|
| 21 |
trust_remote_code=True,
|
| 22 |
)
|
| 23 |
+
tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-2b-jpn-it")
|
| 24 |
upsampler = pipeline(
|
| 25 |
"text-generation",
|
| 26 |
model=model,
|