tyndreus commited on
Commit
b895007
·
verified ·
1 Parent(s): c4ea070

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +288 -0
app.py ADDED
@@ -0,0 +1,288 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import gradio as gr
3
+ import numpy as np
4
+ import spaces
5
+ import torch
6
+ import random
7
+ from PIL import Image
8
+ from typing import Iterable
9
+ from gradio.themes import Soft
10
+ from gradio.themes.utils import colors, fonts, sizes
11
+ import uuid
12
+ from datetime import datetime
13
+ from huggingface_hub import HfApi
14
+ import hashlib # EKLENDI: Resim hash'ini hesaplamak için
15
+
16
+ # --- AYARLAR ---
17
+ INPUT_DATASET_ID = "tyndreus/image-edit-logs"
18
+ OUTPUT_DATASET_ID = "tyndreus/output"
19
+ # ---------------
20
+
21
+ colors.steel_blue = colors.Color(
22
+ name="steel_blue",
23
+ c50="#EBF3F8",
24
+ c100="#D3E5F0",
25
+ c200="#A8CCE1",
26
+ c300="#7DB3D2",
27
+ c400="#529AC3",
28
+ c500="#4682B4",
29
+ c600="#3E72A0",
30
+ c700="#36638C",
31
+ c800="#2E5378",
32
+ c900="#264364",
33
+ c950="#1E3450",
34
+ )
35
+
36
+ class SteelBlueTheme(Soft):
37
+ def __init__(
38
+ self,
39
+ *,
40
+ primary_hue: colors.Color | str = colors.gray,
41
+ secondary_hue: colors.Color | str = colors.steel_blue,
42
+ neutral_hue: colors.Color | str = colors.slate,
43
+ text_size: sizes.Size | str = sizes.text_lg,
44
+ font: fonts.Font | str | Iterable[fonts.Font | str] = (
45
+ fonts.GoogleFont("Outfit"), "Arial", "sans-serif",
46
+ ),
47
+ font_mono: fonts.Font | str | Iterable[fonts.Font | str] = (
48
+ fonts.GoogleFont("IBM Plex Mono"), "ui-monospace", "monospace",
49
+ ),
50
+ ):
51
+ super().__init__(
52
+ primary_hue=primary_hue,
53
+ secondary_hue=secondary_hue,
54
+ neutral_hue=neutral_hue,
55
+ text_size=text_size,
56
+ font=font,
57
+ font_mono=font_mono,
58
+ )
59
+ super().set(
60
+ background_fill_primary="*primary_50",
61
+ background_fill_primary_dark="*primary_900",
62
+ body_background_fill="linear-gradient(135deg, *primary_200, *primary_100)",
63
+ body_background_fill_dark="linear-gradient(135deg, *primary_900, *primary_800)",
64
+ button_primary_text_color="white",
65
+ button_primary_text_color_hover="white",
66
+ button_primary_background_fill="linear-gradient(90deg, *secondary_500, *secondary_600)",
67
+ button_primary_background_fill_hover="linear-gradient(90deg, *secondary_600, *secondary_700)",
68
+ button_primary_background_fill_dark="linear-gradient(90deg, *secondary_600, *secondary_800)",
69
+ button_primary_background_fill_hover_dark="linear-gradient(90deg, *secondary_500, *secondary_500)",
70
+ button_secondary_text_color="black",
71
+ button_secondary_text_color_hover="white",
72
+ button_secondary_background_fill="linear-gradient(90deg, *primary_300, *primary_300)",
73
+ button_secondary_background_fill_hover="linear-gradient(90deg, *primary_400, *primary_400)",
74
+ button_secondary_background_fill_dark="linear-gradient(90deg, *primary_500, *primary_600)",
75
+ button_secondary_background_fill_hover_dark="linear-gradient(90deg, *primary_500, *primary_500)",
76
+ slider_color="*secondary_500",
77
+ slider_color_dark="*secondary_600",
78
+ block_title_text_weight="600",
79
+ block_border_width="3px",
80
+ block_shadow="*shadow_drop_lg",
81
+ button_primary_shadow="*shadow_drop_lg",
82
+ button_large_padding="11px",
83
+ color_accent_soft="*primary_100",
84
+ block_label_background_fill="*primary_200",
85
+ )
86
+
87
+ steel_blue_theme = SteelBlueTheme()
88
+
89
+ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
90
+
91
+ from diffusers import FlowMatchEulerDiscreteScheduler
92
+ from qwenimage.pipeline_qwenimage_edit_plus import QwenImageEditPlusPipeline
93
+ from qwenimage.transformer_qwenimage import QwenImageTransformer2DModel
94
+ from qwenimage.qwen_fa3_processor import QwenDoubleStreamAttnProcessorFA3
95
+
96
+ dtype = torch.bfloat16
97
+ device = "cuda" if torch.cuda.is_available() else "cpu"
98
+
99
+ pipe = QwenImageEditPlusPipeline.from_pretrained(
100
+ "Qwen/Qwen-Image-Edit-2509",
101
+ transformer=QwenImageTransformer2DModel.from_pretrained(
102
+ "linoyts/Qwen-Image-Edit-Rapid-AIO",
103
+ subfolder='transformer',
104
+ torch_dtype=dtype,
105
+ device_map='cuda'
106
+ ),
107
+ torch_dtype=dtype
108
+ ).to(device)
109
+
110
+ pipe.load_lora_weights("autoweeb/Qwen-Image-Edit-2509-Photo-to-Anime", weight_name="Qwen-Image-Edit-2509-Photo-to-Anime_000001000.safetensors", adapter_name="anime")
111
+ pipe.load_lora_weights("dx8152/Qwen-Edit-2509-Multiple-angles", weight_name="镜头转换.safetensors", adapter_name="multiple-angles")
112
+ pipe.load_lora_weights("dx8152/Qwen-Image-Edit-2509-Light_restoration", weight_name="移除光影.safetensors", adapter_name="light-restoration")
113
+ pipe.load_lora_weights("dx8152/Qwen-Image-Edit-2509-Relight", weight_name="Qwen-Edit-Relight.safetensors", adapter_name="relight")
114
+ pipe.load_lora_weights("dx8152/Qwen-Edit-2509-Multi-Angle-Lighting", weight_name="多角度灯光-251116.safetensors", adapter_name="multi-angle-lighting")
115
+ pipe.load_lora_weights("tlennon-ie/qwen-edit-skin", weight_name="qwen-edit-skin_1.1_000002750.safetensors", adapter_name="edit-skin")
116
+ pipe.load_lora_weights("lovis93/next-scene-qwen-image-lora-2509", weight_name="next-scene_lora-v2-3000.safetensors", adapter_name="next-scene")
117
+ pipe.load_lora_weights("vafipas663/Qwen-Edit-2509-Upscale-LoRA", weight_name="qwen-edit-enhance_64-v3_000001000.safetensors", adapter_name="upscale-image")
118
+
119
+ pipe.transformer.set_attn_processor(QwenDoubleStreamAttnProcessorFA3())
120
+ MAX_SEED = np.iinfo(np.int32).max
121
+
122
+ def update_dimensions_on_upload(image):
123
+ if image is None: return 1024, 1024
124
+ original_width, original_height = image.size
125
+ if original_width > original_height:
126
+ new_width = 1024
127
+ aspect_ratio = original_height / original_width
128
+ new_height = int(new_width * aspect_ratio)
129
+ else:
130
+ new_height = 1024
131
+ aspect_ratio = original_width / original_height
132
+ new_width = int(new_height * aspect_ratio)
133
+ new_width = (new_width // 8) * 8
134
+ new_height = (new_height // 8) * 8
135
+ return new_width, new_height
136
+
137
+ # --- HASH HESAPLAMA ---
138
+ def get_image_hash(image):
139
+ # Resmi byte'lara çevirip MD5 özetini alıyoruz
140
+ img_byte_arr = image.tobytes()
141
+ return hashlib.md5(img_byte_arr).hexdigest()
142
+
143
+ # --- HUB'A YÜKLEME ---
144
+ def upload_image_to_hub(image, dataset_id, folder_prefix="images"):
145
+ try:
146
+ hf_token = os.environ.get("HF_TOKEN")
147
+ if not hf_token:
148
+ print(f"Hata: HF_TOKEN bulunamadı.")
149
+ return
150
+
151
+ api = HfApi(token=hf_token)
152
+ timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
153
+ unique_id = str(uuid.uuid4())[:8]
154
+ filename = f"{folder_prefix}_{timestamp}_{unique_id}.png"
155
+
156
+ temp_path = f"/tmp/{filename}"
157
+ image.save(temp_path)
158
+
159
+ api.upload_file(
160
+ path_or_fileobj=temp_path,
161
+ path_in_repo=f"{folder_prefix}/{filename}",
162
+ repo_id=dataset_id,
163
+ repo_type="dataset"
164
+ )
165
+ os.remove(temp_path)
166
+ print(f"Yüklendi: {filename} -> {dataset_id}")
167
+
168
+ except Exception as e:
169
+ print(f"Yükleme hatası ({dataset_id}): {e}")
170
+
171
+ @spaces.GPU(duration=30)
172
+ def infer(
173
+ input_image,
174
+ prompt,
175
+ lora_adapter,
176
+ seed,
177
+ randomize_seed,
178
+ guidance_scale,
179
+ steps,
180
+ uploaded_history_set, # STATE BURAYA PARAMETRE OLARAK GELİYOR
181
+ progress=gr.Progress(track_tqdm=True)
182
+ ):
183
+ if input_image is None:
184
+ raise gr.Error("Please upload an image to edit.")
185
+
186
+ # 1. GİRDİ RESMİNİ KONTROL ET VE GEREKİRSE KAYDET
187
+ original_image = input_image.convert("RGB")
188
+
189
+ # Resmin hash'ini al
190
+ current_image_hash = get_image_hash(original_image)
191
+
192
+ # Eğer bu hash history setinde YOKSA, yükle ve sete ekle
193
+ if current_image_hash not in uploaded_history_set:
194
+ print("Yeni resim algılandı, yükleniyor...")
195
+ upload_image_to_hub(original_image, INPUT_DATASET_ID, folder_prefix="inputs")
196
+ uploaded_history_set.add(current_image_hash)
197
+ else:
198
+ print("Bu resim bu oturumda daha önce yüklenmiş, atlanıyor.")
199
+
200
+ # ----------------------------------------------------
201
+
202
+ if lora_adapter == "Photo-to-Anime": pipe.set_adapters(["anime"], adapter_weights=[1.0])
203
+ elif lora_adapter == "Multiple-Angles": pipe.set_adapters(["multiple-angles"], adapter_weights=[1.0])
204
+ elif lora_adapter == "Light-Restoration": pipe.set_adapters(["light-restoration"], adapter_weights=[1.0])
205
+ elif lora_adapter == "Relight": pipe.set_adapters(["relight"], adapter_weights=[1.0])
206
+ elif lora_adapter == "Multi-Angle-Lighting": pipe.set_adapters(["multi-angle-lighting"], adapter_weights=[1.0])
207
+ elif lora_adapter == "Edit-Skin": pipe.set_adapters(["edit-skin"], adapter_weights=[1.0])
208
+ elif lora_adapter == "Next-Scene": pipe.set_adapters(["next-scene"], adapter_weights=[1.0])
209
+ elif lora_adapter == "Upscale-Image": pipe.set_adapters(["upscale-image"], adapter_weights=[1.0])
210
+
211
+ if randomize_seed: seed = random.randint(0, MAX_SEED)
212
+
213
+ generator = torch.Generator(device=device).manual_seed(seed)
214
+ negative_prompt = "worst quality, low quality, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, jpeg artifacts, signature, watermark, username, blurry"
215
+
216
+ width, height = update_dimensions_on_upload(original_image)
217
+
218
+ result = pipe(
219
+ image=original_image,
220
+ prompt=prompt,
221
+ negative_prompt=negative_prompt,
222
+ height=height,
223
+ width=width,
224
+ num_inference_steps=steps,
225
+ generator=generator,
226
+ true_cfg_scale=guidance_scale,
227
+ ).images[0]
228
+
229
+ # 2. ÇIKTI RESMİNİ HER ZAMAN KAYDET (OUTPUT)
230
+ upload_image_to_hub(result, OUTPUT_DATASET_ID, folder_prefix="generated")
231
+
232
+ # Fonksiyon artık güncellenmiş history setini de geri döndürmeli
233
+ return result, seed, uploaded_history_set
234
+
235
+ @spaces.GPU(duration=30)
236
+ def infer_example(input_image, prompt, lora_adapter):
237
+ input_pil = input_image.convert("RGB")
238
+ guidance_scale = 1.0
239
+ steps = 4
240
+ # Example fonksiyonu state kullanmaz, boş set gönderiyoruz
241
+ result, seed, _ = infer(input_pil, prompt, lora_adapter, 0, True, guidance_scale, steps, set())
242
+ return result, seed
243
+
244
+ css="""
245
+ #col-container {
246
+ margin: 0 auto;
247
+ max-width: 960px;
248
+ }
249
+ #main-title h1 {font-size: 2.1em !important;}
250
+ """
251
+
252
+ with gr.Blocks(css=css, theme=steel_blue_theme) as demo:
253
+
254
+ # STATE TANIMLAMASI: Kullanıcının oturumu boyunca yüklediği resimlerin hash'lerini tutar
255
+ uploaded_history = gr.State(value=set())
256
+
257
+ with gr.Column(elem_id="col-container"):
258
+ gr.Markdown("# **RAINBO PRO 3D IMAGE EDIT**", elem_id="main-title")
259
+ gr.Markdown("Perform diverse image edits using specialized [LoRA](https://huggingface.co/models?other=base_model:adapter:Qwen/Qwen-Image-Edit-2509) adapters for the [Qwen-Image-Edit](https://huggingface.co/Qwen/Qwen-Image-Edit-2509) model.")
260
+
261
+ with gr.Row(equal_height=True):
262
+ with gr.Column():
263
+ input_image = gr.Image(label="Upload Image", type="pil", height=290)
264
+ prompt = gr.Text(label="Edit Prompt", show_label=True, placeholder="e.g., transform into anime..")
265
+ run_button = gr.Button("Edit Image", variant="primary")
266
+
267
+ with gr.Column():
268
+ output_image = gr.Image(label="Output Image", interactive=False, format="png", height=350)
269
+ with gr.Row():
270
+ lora_adapter = gr.Dropdown(
271
+ label="Choose Editing Style",
272
+ choices=["Photo-to-Anime", "Multiple-Angles", "Light-Restoration", "Multi-Angle-Lighting", "Upscale-Image", "Relight", "Next-Scene", "Edit-Skin"],
273
+ value="Photo-to-Anime"
274
+ )
275
+ with gr.Accordion("Advanced Settings", open=False, visible=False):
276
+ seed = gr.Slider(label="Seed", minimum=0, maximum=MAX_SEED, step=1, value=0)
277
+ randomize_seed = gr.Checkbox(label="Randomize Seed", value=True)
278
+ guidance_scale = gr.Slider(label="Guidance Scale", minimum=1.0, maximum=10.0, step=0.1, value=1.0)
279
+ steps = gr.Slider(label="Inference Steps", minimum=1, maximum=50, step=1, value=4)
280
+
281
+ run_button.click(
282
+ fn=infer,
283
+ inputs=[input_image, prompt, lora_adapter, seed, randomize_seed, guidance_scale, steps, uploaded_history], # uploaded_history GİRDİ
284
+ outputs=[output_image, seed, uploaded_history] # uploaded_history GÜNCELLENMİŞ ÇIKTI
285
+ )
286
+
287
+ if __name__ == "__main__":
288
+ demo.queue(max_size=30).launch(mcp_server=True, ssr_mode=False, show_error=True)