# import torch # from diffusers import FluxKontextPipeline, FluxPipeline # from diffusers.utils import load_image # pipe = FluxKontextPipeline.from_pretrained("/data/xcl/Flux-Kontext/model/FLUX-Kontext-dev", torch_dtype=torch.bfloat16) # # pipe.to("cuda") # pipe.enable_model_cpu_offload() # pipe.image_processor # input_image = load_image("/data/xcl/dataSet/RSICD_1/test_png/7.png") # image = pipe( # image=input_image, # # prompt="Make the paper in the image appear wrinkled and crumpled.", # # prompt="Make the handwriting appear messy and wobbly, as if written by a student in a hurry or with uneven hand pressure. Keep the same content and layout.", # # prompt="Make the text in the image look like it was handwritten perfunctorily on paper.", # # prompt="Reduce overall brightness, add mild shadows, and lower contrast slightly while keeping the handwriting readable.", # # prompt="Change the background to look like aged, yellowed paper with slight stains or discoloration.", # # prompt="Add a few natural-looking coffee stains or water rings to the background", # prompt="Replace planes with trucks.", # guidance_scale=2.5, # height=512, # width=512, # ).images[0] # image.save("junshi.png") # import faulthandler # # 在import之后直接添加以下启用代码即可 # faulthandler.enable() # import torch # from diffusers import FluxKontextPipeline # from diffusers.utils import load_image # pipe = FluxKontextPipeline.from_pretrained("/data/xcl/Flux-Kontext/model/FLUX-Kontext-dev", torch_dtype=torch.bfloat16, device_map="balanced") # # pipe = FluxKontextPipeline.from_pretrained("/data/xcl/model/flux-kontext", torch_dtype=torch.bfloat16) # # pipe.to("cuda") # # pipe.enable_model_cpu_offload() # input_image = load_image("/data/xcl/dataSet/images/8.png") # # input_image = input_image.resize((512, 512)) # image = pipe( # image=input_image, # prompt="Replace the ships with airplanes.", # guidance_scale=2.5, # ).images[0] # image.save("8.png") # import os # import json # from PIL import Image # import torch # from diffusers import FluxKontextPipeline # def process_images(): # # 初始化管道 - 使用 FluxKontextPipeline # pipeline = FluxKontextPipeline.from_pretrained( # "/data/xcl/Flux-Kontext/model/FLUX-Kontext-dev", # torch_dtype=torch.bfloat16, # device_map="balanced" # ) # # 启用 CPU offload 以节省显存 # # pipeline.enable_model_cpu_offload() # print("FluxKontextPipeline loaded with automatic device mapping and CPU offload.") # pipeline.set_progress_bar_config(disable=None) # # 路径配置 # input_dir = "/data/xcl/dataSet/images" # output_dir = "/data/xcl/dataSet/images_entity_kontext" # json_file = "/data/xcl/dataSet/junshi_images_entity.json" # # 创建输出目录 # os.makedirs(output_dir, exist_ok=True) # # 加载JSON文件 # with open(json_file, 'r') as f: # prompt_dict = json.load(f) # print(f"Loaded {len(prompt_dict)} image prompts from JSON file.") # # 处理每张图片 # processed_count = 0 # for filename, prompt in prompt_dict.items(): # input_path = os.path.join(input_dir, filename) # output_path = os.path.join(output_dir, filename) # # 检查输入图片是否存在 # if not os.path.exists(input_path): # print(f"Warning: Image {input_path} not found, skipping...") # continue # # 检查输出是否已存在(避免重复处理) # if os.path.exists(output_path): # print(f"Warning: Output {output_path} already exists, skipping...") # continue # try: # # 加载并处理图片 # image = Image.open(input_path).convert("RGB") # width, height = image.size # # 准备输入参数 - 根据新模型的API调整 # inputs = { # "image": image, # "prompt": prompt, # "guidance_scale": 2.5, # 新模型使用 guidance_scale 而不是 true_cfg_scale # "height": height, # 使用输入图像的高度 # "width": width, # 使用输入图像的宽度 # # "generator": torch.manual_seed(0), # 新模型可能不需要这个参数 # # "negative_prompt": " ", # 新模型可能不需要negative_prompt # # "num_inference_steps": 50, # 新模型可能使用默认步数 # } # # 执行图像编辑 # with torch.inference_mode(): # output = pipeline(**inputs) # output_image = output.images[0] # output_image.save(output_path) # # 清理GPU缓存 # if torch.cuda.is_available(): # torch.cuda.empty_cache() # processed_count += 1 # print(f"Processed {filename} -> {output_path}") # except Exception as e: # print(f"Error processing {filename}: {str(e)}") # # 出错时也清理GPU缓存 # if torch.cuda.is_available(): # torch.cuda.empty_cache() # continue # print(f"Processing completed! Successfully processed {processed_count} images.") # if __name__ == "__main__": # process_images() import os import json from PIL import Image import torch from diffusers import FluxKontextPipeline def resize_image_if_needed(image, max_size=1024): """ 如果图片的最长边超过max_size,则按比例调整大小 """ width, height = image.size max_dimension = max(width, height) if max_dimension <= max_size: return image # 计算新的尺寸,保持宽高比 if width > height: new_width = max_size new_height = int(height * (max_size / width)) else: new_height = max_size new_width = int(width * (max_size / height)) # 使用LANCZOS重采样以获得更好的质量 resized_image = image.resize((new_width, new_height), Image.LANCZOS) print(f"Resized image from {width}x{height} to {new_width}x{new_height}") return resized_image def process_images(): # 初始化管道 - 使用 FluxKontextPipeline pipeline = FluxKontextPipeline.from_pretrained( "/data/xcl/Flux-Kontext/model/FLUX-Kontext-dev", torch_dtype=torch.bfloat16, device_map="balanced" ) # 启用 CPU offload 以节省显存 # pipeline.enable_model_cpu_offload() print("FluxKontextPipeline loaded with automatic device mapping and CPU offload.") pipeline.set_progress_bar_config(disable=None) # 路径配置 input_dir = "/data/xcl/dataSet/images" output_dir = "/data/xcl/dataSet/images_entity_background_kontext" json_file = "/data/xcl/dataSet/junshi_images_entity_background.json" # 创建输出目录 os.makedirs(output_dir, exist_ok=True) # 加载JSON文件 with open(json_file, 'r') as f: prompt_dict = json.load(f) print(f"Loaded {len(prompt_dict)} image prompts from JSON file.") # 处理每张图片 processed_count = 0 for filename, prompt in prompt_dict.items(): input_path = os.path.join(input_dir, filename) output_path = os.path.join(output_dir, filename) # 检查输入图片是否存在 if not os.path.exists(input_path): print(f"Warning: Image {input_path} not found, skipping...") continue # 检查输出是否已存在(避免重复处理) if os.path.exists(output_path): print(f"Warning: Output {output_path} already exists, skipping...") continue try: # 加载图片 image = Image.open(input_path).convert("RGB") original_width, original_height = image.size # 检查并调整图片尺寸 image = resize_image_if_needed(image, max_size=1024) new_width, new_height = image.size # 准备输入参数 - 根据新模型的API调整 inputs = { "image": image, "prompt": prompt, "guidance_scale": 2.5, # 新模型使用 guidance_scale 而不是 true_cfg_scale "height": new_height, # 使用调整后的高度 "width": new_width, # 使用调整后的宽度 # "generator": torch.manual_seed(0), # 新模型可能不需要这个参数 # "negative_prompt": " ", # 新模型可能不需要negative_prompt # "num_inference_steps": 50, # 新模型可能使用默认步数 } # 执行图像编辑 with torch.inference_mode(): output = pipeline(**inputs) output_image = output.images[0] output_image.save(output_path) # 清理GPU缓存 if torch.cuda.is_available(): torch.cuda.empty_cache() processed_count += 1 print(f"Processed {filename} (original: {original_width}x{original_height}, processed: {new_width}x{new_height}) -> {output_path}") except Exception as e: print(f"Error processing {filename}: {str(e)}") # 出错时也清理GPU缓存 if torch.cuda.is_available(): torch.cuda.empty_cache() continue print(f"Processing completed! Successfully processed {processed_count} images.") if __name__ == "__main__": process_images()