nkzlxs commited on
Commit
261d9d4
Β·
verified Β·
1 Parent(s): 18ef2ef

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +34 -28
app.py CHANGED
@@ -4,47 +4,53 @@ from diffusers import FluxPipeline
4
  from safetensors.torch import load_file
5
  import os
6
 
7
- # ==============================
8
  # CONFIG β€” ADD YOUR HF TOKEN HERE
9
- # ==============================
10
- HF_TOKEN = os.getenv('HF_TOKEN') # πŸ”‘ REPLACE WITH YOUR ACTUAL TOKEN
11
  HF_MODEL = "black-forest-labs/FLUX.1-dev"
12
  LORA_FILE = "./lora/20.safetensors"
13
 
14
- # ==============================
15
  # LOAD PIPELINE WITH AUTH
16
- # ==============================
17
- pipe = FluxPipeline.from_pretrained(
18
- HF_MODEL,
19
- torch_dtype=torch.bfloat16,
20
- use_safetensors=True,
21
- use_auth_token=HF_TOKEN, # βœ… AUTHENTICATE HERE
22
- ).to("cuda")
 
 
 
 
23
 
24
- # ==============================
25
  # LOAD LORA
26
- # ==============================
27
  if os.path.exists(LORA_FILE):
28
- lora = load_file(LORA_FILE, device="cuda")
29
- pipe.load_lora_weights(lora)
30
- pipe.fuse_lora(lora_scale=1.0)
 
 
 
 
31
 
32
- # ==============================
33
  # GENERATE
34
- # ==============================
35
  def generate(prompt, seed=42):
 
36
  generator = torch.Generator("cuda").manual_seed(seed)
37
- return pipe(
38
- prompt,
39
- generator=generator,
40
- num_inference_steps=28,
41
- height=1024,
42
- width=1024,
43
- ).images[0]
 
 
 
 
 
 
44
 
45
- # ==============================
46
  # GRADIO
47
- # ==============================
48
  with gr.Blocks() as demo:
49
  gr.Markdown("# 🎨 FLUX.1 + My LoRA")
50
  prompt = gr.Textbox(label="Prompt", value="portrait of san, realistic, 8k")
@@ -53,4 +59,4 @@ with gr.Blocks() as demo:
53
  gr.Button("Generate").click(generate, [prompt, seed], output)
54
 
55
  if __name__ == "__main__":
56
- demo.launch()
 
4
  from safetensors.torch import load_file
5
  import os
6
 
 
7
  # CONFIG β€” ADD YOUR HF TOKEN HERE
8
+ HF_TOKEN = os.getenv('HF_TOKEN')
 
9
  HF_MODEL = "black-forest-labs/FLUX.1-dev"
10
  LORA_FILE = "./lora/20.safetensors"
11
 
 
12
  # LOAD PIPELINE WITH AUTH
13
+ try:
14
+ pipe = FluxPipeline.from_pretrained(
15
+ HF_MODEL,
16
+ torch_dtype=torch.float16, # Change to float16
17
+ use_safetensors=True,
18
+ use_auth_token=HF_TOKEN,
19
+ ).to("cuda")
20
+ print("Model loaded successfully.")
21
+ except Exception as e:
22
+ print(f"Error loading model: {e}")
23
+ exit()
24
 
 
25
  # LOAD LORA
 
26
  if os.path.exists(LORA_FILE):
27
+ try:
28
+ lora = load_file(LORA_FILE, device="cuda")
29
+ pipe.load_lora_weights(lora)
30
+ pipe.fuse_lora(lora_scale=1.0)
31
+ print("LoRA loaded successfully.")
32
+ except Exception as e:
33
+ print(f"Error loading LoRA: {e}")
34
 
 
35
  # GENERATE
 
36
  def generate(prompt, seed=42):
37
+ seed = int(seed)
38
  generator = torch.Generator("cuda").manual_seed(seed)
39
+ try:
40
+ result = pipe(
41
+ prompt,
42
+ generator=generator,
43
+ num_inference_steps=28,
44
+ height=1024,
45
+ width=1024,
46
+ ).images[0]
47
+ print("Image generated successfully.")
48
+ return result
49
+ except Exception as e:
50
+ print(f"Error during image generation: {e}")
51
+ return None
52
 
 
53
  # GRADIO
 
54
  with gr.Blocks() as demo:
55
  gr.Markdown("# 🎨 FLUX.1 + My LoRA")
56
  prompt = gr.Textbox(label="Prompt", value="portrait of san, realistic, 8k")
 
59
  gr.Button("Generate").click(generate, [prompt, seed], output)
60
 
61
  if __name__ == "__main__":
62
+ demo.launch()