brothelsnsprout commited on
Commit
c91289c
·
verified ·
1 Parent(s): dc9ae47

Rename main.py to app.py

Browse files
Files changed (2) hide show
  1. app.py +33 -0
  2. main.py +0 -29
app.py ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import torch
3
+ from PIL import Image
4
+ from diffusers import StableDiffusionImg2ImgPipeline
5
+
6
+ # Load your model (replace with your own model repo)
7
+ model_id = "brothelsnsprout/Training_image_generator"
8
+
9
+ pipe = StableDiffusionImg2ImgPipeline.from_pretrained(
10
+ model_id,
11
+ torch_dtype=torch.float16
12
+ ).to("cuda")
13
+
14
+ # Stylize function
15
+ def stylize_image(image: Image.Image, prompt: str):
16
+ image = image.convert("RGB").resize((512, 512))
17
+ result = pipe(prompt=prompt, image=image, strength=0.8, guidance_scale=7.5)
18
+ return result.images[0]
19
+
20
+ # Gradio Interface
21
+ interface = gr.Interface(
22
+ fn=stylize_image,
23
+ inputs=[
24
+ gr.Image(type="pil", label="Input Image"),
25
+ gr.Textbox(label="Text Prompt (Style Instruction)", placeholder="e.g. in the style of Studio Ghibli")
26
+ ],
27
+ outputs=gr.Image(label="Styled Output"),
28
+ title="Image-to-Image Styler",
29
+ description="Upload an image and describe how it should be transformed in style.",
30
+ )
31
+
32
+ if __name__ == "__main__":
33
+ interface.launch()
main.py DELETED
@@ -1,29 +0,0 @@
1
- import gradio as gr
2
- import modin.pandas as pd
3
- import torch
4
- import numpy as np
5
- from PIL import Image
6
- from diffusers import DiffusionPipeline
7
- from huggingface_hub import login
8
-
9
- #import os
10
-
11
- #login(token=os.environ.get('HF_KEY'))
12
-
13
- device = "cuda" if torch.cuda.is_available() else "cpu"
14
- pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-refiner-1.0", torch_dtype=torch.float16) if torch.cuda.is_available() else DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-refiner-1.0")
15
- pipe = pipe.to(device)
16
- # ok
17
- def infer(prompt, source_image, negative_prompt, guide, steps, seed, Strength):
18
- seed = int(seed)
19
- generator = torch.Generator(device).manual_seed(seed)
20
- if not isinstance(source_image, Image.Image):
21
- source_image = Image.open(source_image).convert("RGB")
22
- image = pipe(prompt, negative_prompt=negative_prompt, image=source_image, strength=Strength, guidance_scale=guide, num_inference_steps=steps).images[0]
23
- return image
24
-
25
- gr.Interface(fn=infer, inputs=[gr.Text(label="Prompt"), gr.Image(label="Initial Image", type="pil"), gr.Text(label="Prompt"), gr.Slider(2, 15, value = 7, label = 'Guidance Scale'),
26
- gr.Slider(1, 25, value = 10, step = 1, label = 'Number of Iterations'),
27
- gr.Slider(label = "Seed", minimum = 0, maximum = 987654321987654321, step = 1, randomize = True),
28
- gr.Slider(label='Strength', minimum = 0, maximum = 1, step = .05, value = .5)],
29
- outputs='image', title = "Stable Diffusion XL 1.0 Image to Image Pipeline CPU", description = "For more information on Stable Diffusion XL 1.0 see https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0 <br><br>Upload an Image (<b>MUST Be .PNG and 512x512 or 768x768</b>) enter a Prompt, or let it just do its Thing, then click submit. 10 Iterations takes about ~900-1200 seconds currently. For more informationon about Stable Diffusion or Suggestions for prompts, keywords, artists or styles see https://github.com/Maks-s/sd-akashic", article = "CuongTran").queue(max_size=5).launch(share=True)