Spaces:
Runtime error
Runtime error
| # 部署 teed、depth-anything | |
| # 腐蚀算法 | |
| # 读取图片 | |
| # 输出图片 | |
| # 使用 depth-anything + teed 生成外轮廓 | |
| # 使用 teed + 腐蚀算法 生成内边缘 | |
| import zipfile | |
| from PIL import Image | |
| import cv2 | |
| import numpy as np | |
| import gradio as gr | |
| import os | |
| import torch | |
| import torch.nn.functional as F | |
| from torchvision.transforms import Compose | |
| from tqdm import tqdm | |
| import TEED.main as teed | |
| from TEED.main import parse_args | |
| import logging | |
| from depthAnything.depth_anything.dpt import DepthAnything | |
| from depthAnything.depth_anything.util.transform import Resize, NormalizeImage, PrepareForNet | |
| import shutil | |
| def multiply_blend(image1, image2): | |
| # 将图片转换为浮点数,方便计算 | |
| # Ensure image2 has the same shape as image1 | |
| image2 = np.stack((image2,) * 3, axis=-1) | |
| # Perform the blending | |
| multiplied = np.multiply(image1 / 255.0, image2 / 255.0) * 255.0 | |
| return multiplied.astype(np.uint8) | |
| # Example usage | |
| image1 = np.random.randint(0, 256, (717, 790, 3), dtype=np.uint8) | |
| image2 = np.random.randint(0, 256, (717, 790), dtype=np.uint8) | |
| result = multiply_blend(image1, image2) | |
| print(result.shape) # Should be (717, 790, 3) | |
| def screen_blend(image1, image2): | |
| # 将图片转换为浮点数,方便计算 | |
| image1 = image1.astype(float) | |
| image2 = image2.astype(float) | |
| # 执行滤色操作 | |
| screened = 1 - (1 - image1 / 255) * (1 - image2 / 255) * 255 | |
| # 将结果转换回uint8 | |
| result = np.clip(screened, 0, 255).astype('uint8') | |
| return result | |
| def erosion(img, kernel_size=3, iterations=1, dilate=False): | |
| # 灰度化 | |
| if len(img.shape) == 3: | |
| img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) | |
| # # 二值化 | |
| # _, img = cv2.threshold(img, 200, 255, cv2.THRESH_BINARY) | |
| # 腐蚀 | |
| kernel = np.ones((kernel_size, kernel_size), np.uint8) | |
| if dilate: | |
| img = cv2.dilate(img, kernel, iterations=iterations) | |
| else: | |
| img = cv2.erode(img, kernel, iterations=iterations) | |
| return img | |
| def erosion_img_from_path(img_path, output_dir='./output/erosion_img', kernel_size=3, iterations=1, dilate=False): | |
| # 读取图片 | |
| if os.path.isfile(img_path): | |
| name, extension = os.path.splitext(img_path) | |
| if extension: | |
| if extension.lower() == 'txt': | |
| with open(img_path, 'r', encoding='utf-8') as f: | |
| filenames = f.read().splitlines() | |
| elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif']: | |
| filenames = [img_path] | |
| else: | |
| filenames = os.listdir(img_path) | |
| filenames = [os.path.join(img_path, filename) for filename in filenames if | |
| not filename.startswith('.') and filename.lower().endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames.sort() | |
| os.makedirs(output_dir, exist_ok=True) | |
| for filename in tqdm(filenames): | |
| img = cv2.imread(filename) | |
| img = erosion(img, kernel_size, iterations, dilate) | |
| cv2.imwrite(os.path.join(output_dir, os.path.basename(filename)), img) | |
| def copy_file(src, dest): | |
| # 移动文件 | |
| source = src | |
| destination = dest | |
| try: | |
| shutil.copy(source, destination) | |
| except IOError as e: | |
| print("Unable to copy file. %s" % e) | |
| def guassian_blur_path(img_path, output_dir='./output/guassian_blur', kernel_size=3, sigmaX=0): | |
| # 读取图片 | |
| if os.path.isfile(img_path): | |
| name, extension = os.path.splitext(img_path) | |
| if extension: | |
| if extension.lower() == 'txt': | |
| with open(img_path, 'r', encoding='utf-8') as f: | |
| filenames = f.read().splitlines() | |
| elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif']: | |
| filenames = [img_path] | |
| else: | |
| filenames = os.listdir(img_path) | |
| filenames = [os.path.join(img_path, filename) for filename in filenames if | |
| not filename.startswith('.') and filename.lower().endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames.sort() | |
| os.makedirs(output_dir, exist_ok=True) | |
| for filename in tqdm(filenames): | |
| img = cv2.imread(filename) | |
| img = cv2.GaussianBlur(img, (kernel_size, kernel_size), sigmaX) | |
| cv2.imwrite(os.path.join(output_dir, os.path.basename(filename)), img) | |
| def depth_anything(img_path='./input', outdir='./output/depth_anything', encoder='vitl', pred_only=True, | |
| grayscale=True): | |
| # parser = argparse.ArgumentParser() | |
| # parser.add_argument('--img-path', type=str) | |
| # parser.add_argument('--outdir', type=str, default='./vis_depth') | |
| # parser.add_argument('--encoder', type=str, default='vitl', choices=['vits', 'vitb', 'vitl']) | |
| # parser.add_argument('--pred-only', dest='pred_only', action='store_true', help='only display the prediction') | |
| # parser.add_argument('--grayscale', dest='grayscale', action='store_true', help='do not apply colorful palette') | |
| # args = parser.parse_args() | |
| margin_width = 50 | |
| caption_height = 60 | |
| font = cv2.FONT_HERSHEY_SIMPLEX | |
| font_scale = 1 | |
| font_thickness = 2 | |
| DEVICE = 'cuda' if torch.cuda.is_available() else 'cpu' | |
| model_configs = { | |
| 'vitl': {'encoder': 'vitl', 'features': 256, 'out_channels': [256, 512, 1024, 1024]}, | |
| 'vitb': {'encoder': 'vitb', 'features': 128, 'out_channels': [96, 192, 384, 768]}, | |
| 'vits': {'encoder': 'vits', 'features': 64, 'out_channels': [48, 96, 192, 384]} | |
| } | |
| depth_anything = DepthAnything(model_configs[encoder]) | |
| depth_anything.load_state_dict(torch.load('./checkpoints/depth_anything_{}14.pth'.format(encoder))) | |
| depth_anything = depth_anything.to(DEVICE).eval() | |
| total_params = sum(param.numel() for param in depth_anything.parameters()) | |
| print('Total parameters: {:.2f}M'.format(total_params / 1e6)) | |
| transform = Compose([ | |
| Resize( | |
| width=518, | |
| height=518, | |
| resize_target=False, | |
| keep_aspect_ratio=True, | |
| ensure_multiple_of=14, | |
| resize_method='lower_bound', | |
| image_interpolation_method=cv2.INTER_CUBIC, | |
| ), | |
| NormalizeImage(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), | |
| PrepareForNet(), | |
| ]) | |
| if os.path.isfile(img_path): | |
| name, extension = os.path.splitext(img_path) | |
| if extension: | |
| if extension.lower() == 'txt': | |
| with open(img_path, 'r', encoding='utf-8') as f: | |
| filenames = f.read().splitlines() | |
| elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif']: | |
| filenames = [img_path] | |
| else: | |
| filenames = os.listdir(img_path) | |
| filenames = [os.path.join(img_path, filename) for filename in filenames if | |
| not filename.startswith('.') and filename.lower().endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames.sort() | |
| os.makedirs(outdir, exist_ok=True) | |
| for filename in tqdm(filenames): | |
| raw_image = cv2.imread(filename) | |
| image = cv2.cvtColor(raw_image, cv2.COLOR_BGR2RGB) / 255.0 | |
| h, w = image.shape[:2] | |
| image = transform({'image': image})['image'] | |
| image = torch.from_numpy(image).unsqueeze(0).to(DEVICE) | |
| with torch.no_grad(): | |
| depth = depth_anything(image) | |
| depth = F.interpolate(depth[None], (h, w), mode='bilinear', align_corners=False)[0, 0] | |
| depth = (depth - depth.min()) / (depth.max() - depth.min()) * 255.0 | |
| depth = depth.cpu().numpy().astype(np.uint8) | |
| if grayscale: | |
| depth = np.repeat(depth[..., np.newaxis], 3, axis=-1) | |
| else: | |
| depth = cv2.applyColorMap(depth, cv2.COLORMAP_INFERNO) | |
| filename = os.path.basename(filename) | |
| if pred_only: | |
| cv2.imwrite(os.path.join(outdir, filename[:filename.rfind('.')] + '_depth.png'), depth) | |
| else: | |
| split_region = np.ones((raw_image.shape[0], margin_width, 3), dtype=np.uint8) * 255 | |
| combined_results = cv2.hconcat([raw_image, split_region, depth]) | |
| caption_space = np.ones((caption_height, combined_results.shape[1], 3), dtype=np.uint8) * 255 | |
| captions = ['Raw image', 'Depth Anything'] | |
| segment_width = w + margin_width | |
| for i, caption in enumerate(captions): | |
| # Calculate text size | |
| text_size = cv2.getTextSize(caption, font, font_scale, font_thickness)[0] | |
| # Calculate x-coordinate to center the text | |
| text_x = int((segment_width * i) + (w - text_size[0]) / 2) | |
| # Add text caption | |
| cv2.putText(caption_space, caption, (text_x, 40), font, font_scale, (0, 0, 0), font_thickness) | |
| final_result = cv2.vconcat([caption_space, combined_results]) | |
| cv2.imwrite(os.path.join(outdir, filename[:filename.rfind('.')] + '_img_depth.png'), final_result) | |
| def teed_imgs(img_path='./input', outdir='./output/teed_imgs', gaussianBlur=[0, 3, 0]): | |
| args, train_info = parse_args(is_testing=True, pl_opt_dir=outdir) | |
| os.makedirs('teed_tmp', exist_ok=True) | |
| if os.path.isfile(img_path): | |
| name, extension = os.path.splitext(img_path) | |
| if extension: | |
| if extension.lower() == 'txt': | |
| with open(img_path, 'r', encoding='utf-8') as f: | |
| filenames = f.read().splitlines() | |
| elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif']: | |
| filenames = [img_path] | |
| else: | |
| filenames = os.listdir(img_path) | |
| filenames = [os.path.join(img_path, filename) for filename in filenames if | |
| not filename.startswith('.') and filename.lower().endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames.sort() | |
| for filename in tqdm(filenames): | |
| if gaussianBlur[0] != 0: | |
| img = cv2.imread(filename) | |
| img = cv2.GaussianBlur(img, (gaussianBlur[1], gaussianBlur[1]), gaussianBlur[2]) | |
| cv2.imwrite(os.path.join('teed_tmp', os.path.basename(filename)), img) | |
| else: | |
| copy_file(filename, 'teed_tmp') | |
| teed.main(args, train_info) | |
| shutil.rmtree('teed_tmp') | |
| def merge_2_images(img1, img2, mode, erosion_para=[[0, 0], [0, 0]], dilate=[0, 0]): # 将 img1 合并至 img2,调整大小与 img2 相同 | |
| img1 = cv2.imread(img1) | |
| img2 = cv2.imread(img2) | |
| img1 = cv2.resize(img1, (img2.shape[1], img2.shape[0])) | |
| if erosion_para[0][1] != 0: | |
| img1 = erosion(img1, erosion_para[0][0], erosion_para[0][1], dilate[0]) | |
| if erosion_para[1][1] != 0: | |
| img2 = erosion(img2, erosion_para[1][0], erosion_para[1][1], dilate[1]) | |
| if mode == 'multiply': | |
| return multiply_blend(img1, img2) | |
| elif mode == 'screen': | |
| return screen_blend(img1, img2) | |
| def merge_images_in_2_folder(folder1, folder2, outdir, suffix_need_remove=None, suffix_floder=0, mode='multiply', | |
| erosion_para=[[0, 0], [0, 0]], | |
| dilate=[0, 0]): # 将 folder1 和 folder2 中的图片合并,可选是否移除某文件夹后缀,可选腐蚀参数[kernel_size,iterations] | |
| os.makedirs(outdir, exist_ok=True) | |
| name_extension_pairs_folder1 = [os.path.splitext(filename) for filename in os.listdir(folder1) if filename.endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames_noext_folder1, extensions_folder1 = zip(*name_extension_pairs_folder1) | |
| name_extension_pairs_folder2 = [os.path.splitext(filename) for filename in os.listdir(folder2) if filename.endswith( | |
| ('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| filenames_noext_folder2, extensions_folder2 = zip(*name_extension_pairs_folder2) | |
| if suffix_need_remove: | |
| if suffix_floder == 0: | |
| filenames_raw = list(filenames_noext_folder1).copy() | |
| filenames_noext_folder1 = [ | |
| filename[:-len(suffix_need_remove)] + filename[-len(suffix_need_remove):].replace(suffix_need_remove, | |
| '') for filename in | |
| filenames_noext_folder1] | |
| if suffix_floder == 1: | |
| filenames_raw = list(filenames_noext_folder2).copy() | |
| filenames_noext_folder2 = [ | |
| filename[:-len(suffix_need_remove)] + filename[-len(suffix_need_remove):].replace(suffix_need_remove, | |
| '') for filename in | |
| filenames_noext_folder2] | |
| for index, filename in enumerate(filenames_noext_folder1): | |
| if filename in filenames_noext_folder2: | |
| print(filename) | |
| if suffix_need_remove: | |
| if suffix_floder == 0: | |
| img1 = os.path.join(folder1, filenames_raw[index] + extensions_folder1[index]) | |
| img2 = os.path.join(folder2, filename + extensions_folder2[filenames_noext_folder2.index(filename)]) | |
| if suffix_floder == 1: | |
| img1 = os.path.join(folder1, filename + extensions_folder1[index]) | |
| img2 = os.path.join(folder2, | |
| filenames_raw[filenames_noext_folder2.index(filename)] + extensions_folder2[ | |
| filenames_noext_folder2.index(filename)]) | |
| else: | |
| img1 = os.path.join(folder1, filename + extensions_folder1[index]) | |
| img2 = os.path.join(folder2, filename + extensions_folder2[filenames_noext_folder2.index(filename)]) | |
| result = merge_2_images(img1, img2, mode, erosion_para, dilate) | |
| cv2.imwrite(os.path.join(outdir, filename + extensions_folder1[index]), result) | |
| def invert_image(image): | |
| # 将图片从BGR转为灰度图 | |
| gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) | |
| # 对灰度图进行反转 | |
| inverted_image = cv2.bitwise_not(gray_image) | |
| # 将反转后的灰度图转换回BGR格式 | |
| inverted_image_bgr = cv2.cvtColor(inverted_image, cv2.COLOR_GRAY2BGR) | |
| return inverted_image_bgr | |
| def process_images(input_folder='./output/merged_imgs'): | |
| output_folder = os.path.join(os.path.dirname(input_folder), 'output_invert') | |
| os.makedirs(output_folder, exist_ok=True) | |
| # 获取输入文件夹中的所有图片文件 | |
| image_files = [f for f in os.listdir(input_folder) if | |
| f.lower().endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))] | |
| for image_file in tqdm(image_files): | |
| image_path = os.path.join(input_folder, image_file) | |
| try: | |
| # 使用PIL库读取图像 | |
| with Image.open(image_path) as img: | |
| image = np.array(img.convert('RGB'))[:, :, ::-1].copy() | |
| if image is not None: | |
| # 翻转图片 | |
| inverted_image = invert_image(image) | |
| # 保存翻转后的图片到输出文件夹 | |
| output_path = os.path.join(output_folder, image_file) | |
| cv2.imwrite(output_path, inverted_image) | |
| else: | |
| raise ValueError(f"Failed to read image: {image_file}") | |
| except Exception as e: | |
| print(f"Error processing file {image_file}: {e}") | |
| def process_line(input_files): | |
| try: | |
| # 创建临时输出文件夹 | |
| output_folder = "temp_output" | |
| os.makedirs(output_folder, exist_ok=True) | |
| # 存储处理后的图片路径 | |
| processed_images = [] | |
| # 遍历所有输入文件 | |
| for img_path in input_files: | |
| img_path = img_path.name # 获取文件路径 | |
| # 处理图片的文件夹 | |
| depth_folder = os.path.join(output_folder, "depth_anything") | |
| teed_folder = os.path.join(output_folder, "teed_imgs") | |
| dp_teed_folder = os.path.join(output_folder, "dp_teed_imgs") | |
| merged_folder = os.path.join(output_folder, "merged_imgs") | |
| # 创建每个处理步骤的文件夹 | |
| os.makedirs(depth_folder, exist_ok=True) | |
| os.makedirs(teed_folder, exist_ok=True) | |
| os.makedirs(dp_teed_folder, exist_ok=True) | |
| os.makedirs(merged_folder, exist_ok=True) | |
| # 调用处理函数 | |
| depth_anything(img_path, depth_folder) | |
| teed_imgs(img_path, teed_folder, [1, 7, 2]) | |
| teed_imgs(depth_folder, dp_teed_folder, [0, 7, 2]) | |
| merge_images_in_2_folder(teed_folder, dp_teed_folder, merged_folder, '_depth', 1, 'multiply', [[2, 0], [2, 1]], [1, 0]) | |
| process_images(merged_folder) | |
| # 创建压缩包 | |
| zip_file_path = os.path.join(output_folder, "processed_images.zip") | |
| with zipfile.ZipFile(zip_file_path, 'w') as zipf: | |
| # 将每个步骤的文件夹添加到压缩包中 | |
| for folder in [depth_folder, teed_folder, dp_teed_folder, merged_folder]: | |
| for root, _, files in os.walk(folder): | |
| for file in files: | |
| zipf.write(os.path.join(root, file), os.path.relpath(os.path.join(root, file), output_folder)) | |
| return [zip_file_path], "" # 返回压缩包路径和空错误信息 | |
| except Exception as e: | |
| return [], f"发生错误: {str(e)}" # 返回空图片和错误信息 | |
| def launch_interface(): | |
| with gr.Blocks() as demo: | |
| # 允许用户选择多张图片 | |
| input_files = gr.File(label="选择输入图片", file_count="multiple", type="filepath") | |
| submit_button = gr.Button("开始处理") | |
| # 显示处理后的文件下载链接 | |
| output_file = gr.File(label="下载处理后的文件") | |
| # 显示错误信息 | |
| error_text = gr.Textbox(label="错误信息", interactive=False, visible=False) | |
| # 点击按钮时调用 process_line 函数 | |
| submit_button.click(process_line, inputs=[input_files], outputs=[output_file, error_text]) | |
| demo.launch(share=True) | |
| if __name__ == "__main__": | |
| launch_interface() | |