moreover commited on
Commit
c7b534d
·
verified ·
1 Parent(s): 9c6040b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +21 -14
app.py CHANGED
@@ -6,15 +6,24 @@ import spaces #[uncomment to use ZeroGPU]
6
  from diffusers import DiffusionPipeline
7
  import torch
8
 
 
9
  device = "cuda" if torch.cuda.is_available() else "cpu"
 
10
  MODEL_OPTIONS = [
11
  ("stabilityai/sdxl-turbo", "SDXL Turbo (Быстро)"),
12
  ("hakurei/waifu-diffusion", "Что-то альтернативное"),
13
  ]
14
  DEFAULT_MODEL_ID = "stabilityai/sdxl-turbo"
15
- # model_repo_id = "stabilityai/sdxl-turbo" # Replace to the model you would like to use
16
 
17
- PIPELINES = {}
 
 
 
 
 
 
 
 
18
 
19
  def load_pipelines():
20
  # SDXL Turbo
@@ -29,17 +38,8 @@ def load_pipelines():
29
  pipe = pipe.to(device)
30
  PIPELINES[mid] = pipe
31
 
32
-
33
- if torch.cuda.is_available():
34
- torch_dtype = torch.float16
35
- else:
36
- torch_dtype = torch.float32
37
-
38
  load_pipelines()
39
 
40
- MAX_SEED = np.iinfo(np.int32).max
41
- MAX_IMAGE_SIZE = 1024
42
-
43
 
44
  @spaces.GPU #[uncomment to use ZeroGPU]
45
  def infer(
@@ -90,7 +90,12 @@ css = """
90
  with gr.Blocks(css=css) as demo:
91
  with gr.Column(elem_id="col-container"):
92
  gr.Markdown(" # Text-to-Image Gradio Template")
93
-
 
 
 
 
 
94
  with gr.Row():
95
  prompt = gr.Text(
96
  label="Prompt",
@@ -109,7 +114,8 @@ with gr.Blocks(css=css) as demo:
109
  label="Negative prompt",
110
  max_lines=1,
111
  placeholder="Enter a negative prompt",
112
- visible=False,
 
113
  )
114
 
115
  seed = gr.Slider(
@@ -161,6 +167,7 @@ with gr.Blocks(css=css) as demo:
161
  triggers=[run_button.click, prompt.submit],
162
  fn=infer,
163
  inputs=[
 
164
  prompt,
165
  negative_prompt,
166
  seed,
@@ -174,4 +181,4 @@ with gr.Blocks(css=css) as demo:
174
  )
175
 
176
  if __name__ == "__main__":
177
- demo.launch()
 
6
  from diffusers import DiffusionPipeline
7
  import torch
8
 
9
+
10
  device = "cuda" if torch.cuda.is_available() else "cpu"
11
+
12
  MODEL_OPTIONS = [
13
  ("stabilityai/sdxl-turbo", "SDXL Turbo (Быстро)"),
14
  ("hakurei/waifu-diffusion", "Что-то альтернативное"),
15
  ]
16
  DEFAULT_MODEL_ID = "stabilityai/sdxl-turbo"
 
17
 
18
+ if torch.cuda.is_available():
19
+ torch_dtype = torch.float16
20
+ else:
21
+ torch_dtype = torch.float32
22
+
23
+ MAX_SEED = np.iinfo(np.int32).max
24
+ MAX_IMAGE_SIZE = 1024
25
+
26
+ PIPELINES = {}
27
 
28
  def load_pipelines():
29
  # SDXL Turbo
 
38
  pipe = pipe.to(device)
39
  PIPELINES[mid] = pipe
40
 
 
 
 
 
 
 
41
  load_pipelines()
42
 
 
 
 
43
 
44
  @spaces.GPU #[uncomment to use ZeroGPU]
45
  def infer(
 
90
  with gr.Blocks(css=css) as demo:
91
  with gr.Column(elem_id="col-container"):
92
  gr.Markdown(" # Text-to-Image Gradio Template")
93
+
94
+ model_id = gr.Dropdown(
95
+ choices=[m[0] for m in MODEL_OPTIONS],
96
+ label="Model",
97
+ value=DEFAULT_MODEL_ID,)
98
+
99
  with gr.Row():
100
  prompt = gr.Text(
101
  label="Prompt",
 
114
  label="Negative prompt",
115
  max_lines=1,
116
  placeholder="Enter a negative prompt",
117
+ visible=True,
118
+ value="dog, cat"
119
  )
120
 
121
  seed = gr.Slider(
 
167
  triggers=[run_button.click, prompt.submit],
168
  fn=infer,
169
  inputs=[
170
+ model_id,
171
  prompt,
172
  negative_prompt,
173
  seed,
 
181
  )
182
 
183
  if __name__ == "__main__":
184
+ demo.launch()