tejani commited on
Commit
3ef8c1c
·
verified ·
1 Parent(s): a44ba0b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +28 -22
app.py CHANGED
@@ -3,7 +3,10 @@ import numpy as np
3
  import random
4
  import os
5
 
 
6
  hf_token = os.getenv("HF_TOKEN")
 
 
7
  interface = gr.load("models/ZB-Tech/Text-to-Image", token=hf_token)
8
 
9
  MAX_SEED = np.iinfo(np.int32).max
@@ -11,29 +14,32 @@ MAX_IMAGE_SIZE = 1024
11
 
12
  def infer(
13
  prompt,
14
- negative_prompt,
15
- seed,
16
  randomize_seed,
17
- width,
18
- height,
19
- guidance_scale,
20
- num_inference_steps,
21
  progress=gr.Progress(track_tqdm=True),
22
  ):
23
  if randomize_seed:
24
  seed = random.randint(0, MAX_SEED)
25
 
26
- # Use the loaded interface to generate the image
27
- # Note: Parameters depend on what the ZB-Tech/Text-to-Image model supports via the API
28
- image = interface(
29
- prompt=prompt,
30
- negative_prompt=negative_prompt if negative_prompt else None,
31
- seed=seed,
32
- width=width,
33
- height=height,
34
- guidance_scale=guidance_scale,
35
- num_inference_steps=num_inference_steps,
36
- )
 
 
 
37
 
38
  return image, seed
39
 
@@ -72,7 +78,7 @@ with gr.Blocks(css=css) as demo:
72
  label="Negative prompt",
73
  max_lines=1,
74
  placeholder="Enter a negative prompt",
75
- visible=True, # Assuming the API supports it
76
  )
77
 
78
  seed = gr.Slider(
@@ -91,7 +97,7 @@ with gr.Blocks(css=css) as demo:
91
  minimum=256,
92
  maximum=MAX_IMAGE_SIZE,
93
  step=32,
94
- value=512, # Reasonable default for API
95
  )
96
 
97
  height = gr.Slider(
@@ -99,7 +105,7 @@ with gr.Blocks(css=css) as demo:
99
  minimum=256,
100
  maximum=MAX_IMAGE_SIZE,
101
  step=32,
102
- value=512, # Reasonable default for API
103
  )
104
 
105
  with gr.Row():
@@ -108,7 +114,7 @@ with gr.Blocks(css=css) as demo:
108
  minimum=0.0,
109
  maximum=10.0,
110
  step=0.1,
111
- value=1.0, # Default may vary by model
112
  )
113
 
114
  num_inference_steps = gr.Slider(
@@ -116,7 +122,7 @@ with gr.Blocks(css=css) as demo:
116
  minimum=1,
117
  maximum=50,
118
  step=1,
119
- value=10, # Default may vary by model
120
  )
121
 
122
  gr.Examples(examples=examples, inputs=[prompt])
 
3
  import random
4
  import os
5
 
6
+ # Fetch the API token from environment variable
7
  hf_token = os.getenv("HF_TOKEN")
8
+
9
+ # Load the model from Hugging Face Inference API
10
  interface = gr.load("models/ZB-Tech/Text-to-Image", token=hf_token)
11
 
12
  MAX_SEED = np.iinfo(np.int32).max
 
14
 
15
  def infer(
16
  prompt,
17
+ negative_prompt, # May not be supported
18
+ seed, # May not be supported
19
  randomize_seed,
20
+ width, # May not be supported
21
+ height, # May not be supported
22
+ guidance_scale, # May not be supported
23
+ num_inference_steps, # May not be supported
24
  progress=gr.Progress(track_tqdm=True),
25
  ):
26
  if randomize_seed:
27
  seed = random.randint(0, MAX_SEED)
28
 
29
+ # Minimal call to the interface with just the prompt
30
+ # Add other parameters only if the model supports them
31
+ image = interface(prompt=prompt)
32
+
33
+ # If the model supports additional parameters, you can uncomment and test:
34
+ # image = interface(
35
+ # prompt=prompt,
36
+ # negative_prompt=negative_prompt if negative_prompt else None,
37
+ # seed=seed,
38
+ # width=width,
39
+ # height=height,
40
+ # guidance_scale=guidance_scale,
41
+ # num_inference_steps=num_inference_steps,
42
+ # )
43
 
44
  return image, seed
45
 
 
78
  label="Negative prompt",
79
  max_lines=1,
80
  placeholder="Enter a negative prompt",
81
+ visible=True,
82
  )
83
 
84
  seed = gr.Slider(
 
97
  minimum=256,
98
  maximum=MAX_IMAGE_SIZE,
99
  step=32,
100
+ value=512,
101
  )
102
 
103
  height = gr.Slider(
 
105
  minimum=256,
106
  maximum=MAX_IMAGE_SIZE,
107
  step=32,
108
+ value=512,
109
  )
110
 
111
  with gr.Row():
 
114
  minimum=0.0,
115
  maximum=10.0,
116
  step=0.1,
117
+ value=1.0,
118
  )
119
 
120
  num_inference_steps = gr.Slider(
 
122
  minimum=1,
123
  maximum=50,
124
  step=1,
125
+ value=10,
126
  )
127
 
128
  gr.Examples(examples=examples, inputs=[prompt])