Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -15,13 +15,13 @@ repo = "stabilityai/stable-diffusion-3.5-large"
|
|
| 15 |
t2i = StableDiffusion3Pipeline.from_pretrained(repo, torch_dtype=torch.bfloat16, token=os.environ["TOKEN"]).to(device)
|
| 16 |
|
| 17 |
model = AutoModelForCausalLM.from_pretrained(
|
| 18 |
-
"microsoft/Phi-3-mini-instruct",
|
| 19 |
device_map="cuda",
|
| 20 |
torch_dtype=torch.bfloat16,
|
| 21 |
trust_remote_code=True,
|
| 22 |
token=os.environ["TOKEN"]
|
| 23 |
)
|
| 24 |
-
tokenizer = AutoTokenizer.from_pretrained("microsoft/Phi-3-mini-instruct", token=os.environ["TOKEN"])
|
| 25 |
upsampler = pipeline(
|
| 26 |
"text-generation",
|
| 27 |
model=model,
|
|
|
|
| 15 |
t2i = StableDiffusion3Pipeline.from_pretrained(repo, torch_dtype=torch.bfloat16, token=os.environ["TOKEN"]).to(device)
|
| 16 |
|
| 17 |
model = AutoModelForCausalLM.from_pretrained(
|
| 18 |
+
"microsoft/Phi-3-mini-4k-instruct",
|
| 19 |
device_map="cuda",
|
| 20 |
torch_dtype=torch.bfloat16,
|
| 21 |
trust_remote_code=True,
|
| 22 |
token=os.environ["TOKEN"]
|
| 23 |
)
|
| 24 |
+
tokenizer = AutoTokenizer.from_pretrained("microsoft/Phi-3-mini-4k-instruct", token=os.environ["TOKEN"])
|
| 25 |
upsampler = pipeline(
|
| 26 |
"text-generation",
|
| 27 |
model=model,
|