File size: 16,136 Bytes
5cbca7c
 
 
 
 
 
 
c4503c1
5cbca7c
 
 
 
 
 
 
 
 
3b1af3a
5cbca7c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3591a7e
 
5cbca7c
3591a7e
116c9ab
5cbca7c
116c9ab
5cbca7c
3591a7e
116c9ab
5cbca7c
116c9ab
3591a7e
116c9ab
3591a7e
116c9ab
5cbca7c
3591a7e
5cbca7c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
440a7b8
3591a7e
440a7b8
3591a7e
 
440a7b8
3591a7e
 
32dfe62
3591a7e
 
 
 
 
 
 
 
 
 
 
 
51f183c
5cbca7c
 
c4503c1
 
a1fbd3e
c4503c1
1784a51
 
3591a7e
1784a51
c4503c1
1784a51
 
c4503c1
 
 
 
0a78718
 
c4503c1
 
 
c0e34ca
c4503c1
3591a7e
c4503c1
 
0a78718
c4503c1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
# 部署 teed、depth-anything
# 腐蚀算法
# 读取图片
# 输出图片
# 使用 depth-anything + teed 生成外轮廓
# 使用 teed + 腐蚀算法 生成内边缘
from PIL import Image

import cv2
import numpy as np
import os
import torch
import torch.nn.functional as F
from torchvision.transforms import Compose
from tqdm import tqdm
import TEED.main as teed
from TEED.main import parse_args

from depthAnything.depth_anything.dpt import DepthAnything
from depthAnything.depth_anything.util.transform import Resize, NormalizeImage, PrepareForNet
import shutil


def multiply_blend(image1, image2):
    # 将图片转换为浮点数,方便计算
    # Ensure image2 has the same shape as image1
    image2 = np.stack((image2,) * 3, axis=-1)
    # Perform the blending
    multiplied = np.multiply(image1 / 255.0, image2 / 255.0) * 255.0
    return multiplied.astype(np.uint8)

    # Example usage


image1 = np.random.randint(0, 256, (717, 790, 3), dtype=np.uint8)
image2 = np.random.randint(0, 256, (717, 790), dtype=np.uint8)

result = multiply_blend(image1, image2)
print(result.shape)  # Should be (717, 790, 3)

def screen_blend(image1, image2):
    # 将图片转换为浮点数,方便计算
    image1 = image1.astype(float)
    image2 = image2.astype(float)

    # 执行滤色操作
    screened = 1 - (1 - image1 / 255) * (1 - image2 / 255) * 255

    # 将结果转换回uint8
    result = np.clip(screened, 0, 255).astype('uint8')
    return result

def erosion(img, kernel_size = 3, iterations = 1, dilate = False):
    
    # 灰度化
    if len(img.shape) == 3:
        img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
    
    # # 二值化
    # _, img = cv2.threshold(img, 200, 255, cv2.THRESH_BINARY)

    # 腐蚀
    kernel = np.ones((kernel_size, kernel_size), np.uint8)
    if dilate:
        img = cv2.dilate(img, kernel, iterations=iterations)
    else:
        img = cv2.erode(img, kernel, iterations=iterations)

    return img

def erosion_img_from_path(img_path, output_dir = './output/erosion_img', kernel_size = 3, iterations = 1, dilate = False):
    # 读取图片
    if os.path.isfile(img_path):
        name, extension = os.path.splitext(img_path)
        if extension:
            if extension.lower() == 'txt':
                with open(img_path, 'r',encoding= 'utf-8') as f:
                    filenames = f.read().splitlines()
            elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif']:
                filenames = [img_path]
    else:
        filenames = os.listdir(img_path)
        filenames = [os.path.join(img_path, filename) for filename in filenames if not filename.startswith('.') and filename.lower().endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif'))]
        filenames.sort()
    
    os.makedirs(output_dir, exist_ok=True)
    
    for filename in tqdm(filenames):
        img = cv2.imread(filename)
        img = erosion(img, kernel_size, iterations, dilate)
        cv2.imwrite(os.path.join(output_dir, os.path.basename(filename)), img)


def copy_file(src, dest):
    # 移动文件
    source = src
    destination = dest
    try:
        shutil.copy(source, destination)
    except IOError as e:
        print("Unable to copy file. %s" % e)


def guassian_blur_path(img_path, output_dir = './output/guassian_blur', kernel_size = 3, sigmaX = 0):
    # 读取图片
    if os.path.isfile(img_path):
        name, extension = os.path.splitext(img_path)
        if extension:
            if extension.lower() == 'txt':
                with open(img_path, 'r',encoding= 'utf-8') as f:
                    filenames = f.read().splitlines()
            elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif']:
                filenames = [img_path]
    else:
        filenames = os.listdir(img_path)
        filenames = [os.path.join(img_path, filename) for filename in filenames if not filename.startswith('.') and filename.lower().endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif'))]
        filenames.sort()
    
    os.makedirs(output_dir, exist_ok=True)
    
    for filename in tqdm(filenames):
        img = cv2.imread(filename)
        img = cv2.GaussianBlur(img, (kernel_size,kernel_size), sigmaX)
        cv2.imwrite(os.path.join(output_dir, os.path.basename(filename)), img)

def depth_anything(img_path = './input', outdir = './output/depth_anything', encoder = 'vitl', pred_only = True, grayscale = True):
    # parser = argparse.ArgumentParser()
    # parser.add_argument('--img-path', type=str)
    # parser.add_argument('--outdir', type=str, default='./vis_depth')
    # parser.add_argument('--encoder', type=str, default='vitl', choices=['vits', 'vitb', 'vitl'])
    
    # parser.add_argument('--pred-only', dest='pred_only', action='store_true', help='only display the prediction')
    # parser.add_argument('--grayscale', dest='grayscale', action='store_true', help='do not apply colorful palette')
    
    # args = parser.parse_args()
    
    margin_width = 50
    caption_height = 60
    
    font = cv2.FONT_HERSHEY_SIMPLEX
    font_scale = 1
    font_thickness = 2
    
    DEVICE = 'cuda' if torch.cuda.is_available() else 'cpu'

    model_configs = {
    'vitl': {'encoder': 'vitl', 'features': 256, 'out_channels': [256, 512, 1024, 1024]},
    'vitb': {'encoder': 'vitb', 'features': 128, 'out_channels': [96, 192, 384, 768]},
    'vits': {'encoder': 'vits', 'features': 64, 'out_channels': [48, 96, 192, 384]}
    }

    depth_anything = DepthAnything(model_configs[encoder])
    depth_anything.load_state_dict(torch.load('./checkpoints/depth_anything_{}14.pth'.format(encoder)))
    depth_anything = depth_anything.to(DEVICE).eval()
    
    total_params = sum(param.numel() for param in depth_anything.parameters())
    print('Total parameters: {:.2f}M'.format(total_params / 1e6))
    
    transform = Compose([
        Resize(
            width=518,
            height=518,
            resize_target=False,
            keep_aspect_ratio=True,
            ensure_multiple_of=14,
            resize_method='lower_bound',
            image_interpolation_method=cv2.INTER_CUBIC,
        ),
        NormalizeImage(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),
        PrepareForNet(),
    ])
    
    if os.path.isfile(img_path):
        name, extension = os.path.splitext(img_path)
        if extension:
            if extension.lower() == 'txt':
                with open(img_path, 'r',encoding= 'utf-8') as f:
                    filenames = f.read().splitlines()
            elif extension.lower() in ['.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif']:
                filenames = [img_path]
    else:
        filenames = os.listdir(img_path)
        filenames = [os.path.join(img_path, filename) for filename in filenames if not filename.startswith('.') and filename.lower().endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif'))]
        filenames.sort()
    
    os.makedirs(outdir, exist_ok=True)
    
    for filename in tqdm(filenames):
        raw_image = cv2.imread(filename)
        image = cv2.cvtColor(raw_image, cv2.COLOR_BGR2RGB) / 255.0
        
        h, w = image.shape[:2]
        
        image = transform({'image': image})['image']
        image = torch.from_numpy(image).unsqueeze(0).to(DEVICE)
        
        with torch.no_grad():
            depth = depth_anything(image)
        
        depth = F.interpolate(depth[None], (h, w), mode='bilinear', align_corners=False)[0, 0]
        depth = (depth - depth.min()) / (depth.max() - depth.min()) * 255.0
        
        depth = depth.cpu().numpy().astype(np.uint8)
        
        if grayscale:
            depth = np.repeat(depth[..., np.newaxis], 3, axis=-1)
        else:
            depth = cv2.applyColorMap(depth, cv2.COLORMAP_INFERNO)
        
        filename = os.path.basename(filename)
        
        if pred_only:
            cv2.imwrite(os.path.join(outdir, filename[:filename.rfind('.')] + '_depth.png'), depth)
        else:
            split_region = np.ones((raw_image.shape[0], margin_width, 3), dtype=np.uint8) * 255
            combined_results = cv2.hconcat([raw_image, split_region, depth])
            
            caption_space = np.ones((caption_height, combined_results.shape[1], 3), dtype=np.uint8) * 255
            captions = ['Raw image', 'Depth Anything']
            segment_width = w + margin_width
            
            for i, caption in enumerate(captions):
                # Calculate text size
                text_size = cv2.getTextSize(caption, font, font_scale, font_thickness)[0]

                # Calculate x-coordinate to center the text
                text_x = int((segment_width * i) + (w - text_size[0]) / 2)

                # Add text caption
                cv2.putText(caption_space, caption, (text_x, 40), font, font_scale, (0, 0, 0), font_thickness)
            
            final_result = cv2.vconcat([caption_space, combined_results])
            
            cv2.imwrite(os.path.join(outdir, filename[:filename.rfind('.')] + '_img_depth.png'), final_result)

def teed_imgs(img_path='./input', outdir='./output/teed_imgs', gaussianBlur=[0, 3, 0]):
    os.makedirs(outdir, exist_ok=True)
    os.makedirs('teed_tmp', exist_ok=True)

    # 处理单个上传的图片
    if os.path.isfile(img_path):
        img = cv2.imread(img_path)
        if gaussianBlur[0] != 0:
            img = cv2.GaussianBlur(img, (gaussianBlur[1], gaussianBlur[1]), gaussianBlur[2])
            cv2.imwrite(os.path.join('teed_tmp', 'temp_image.png'), img)
        else:
            cv2.imwrite(os.path.join('teed_tmp', 'temp_image.png'), img)

    # 使用临时图像文件进行处理
    args, train_info = parse_args(is_testing=True, pl_opt_dir=outdir)
    args.input_val_dir = 'teed_tmp'  # 使用临时目录
    teed.main(args, train_info)

    shutil.rmtree('teed_tmp')

def merge_2_images(img1, img2, mode, erosion_para = [[0,0],[0,0]], dilate = [0,0]): #将 img1 合并至 img2,调整大小与 img2 相同
    img1 = cv2.imread(img1)
    img2 = cv2.imread(img2)
    img1 = cv2.resize(img1, (img2.shape[1], img2.shape[0]))
    if erosion_para[0][1] != 0:
        img1 = erosion(img1, erosion_para[0][0], erosion_para[0][1], dilate[0])
    if erosion_para[1][1] != 0:
        img2 = erosion(img2, erosion_para[1][0], erosion_para[1][1], dilate[1])
    if mode == 'multiply':
        return multiply_blend(img1, img2)
    elif mode == 'screen':
        return screen_blend(img1, img2)
    
def merge_images_in_2_folder(folder1, folder2, outdir, suffix_need_remove = None, suffix_floder = 0 , mode = 'multiply', erosion_para = [[0,0],[0,0]], dilate = [0,0]): #将 folder1 和 folder2 中的图片合并,可选是否移除某文件夹后缀,可选腐蚀参数[kernel_size,iterations]
    os.makedirs(outdir, exist_ok=True)
    name_extension_pairs_folder1 = [os.path.splitext(filename) for filename in os.listdir(folder1) if filename.endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif'))]
    filenames_noext_folder1, extensions_folder1 = zip(*name_extension_pairs_folder1)
    name_extension_pairs_folder2 = [os.path.splitext(filename) for filename in os.listdir(folder2) if filename.endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp','tif'))]
    filenames_noext_folder2, extensions_folder2 = zip(*name_extension_pairs_folder2)
    if suffix_need_remove:
        if suffix_floder == 0:
            filenames_raw = list(filenames_noext_folder1).copy()
            filenames_noext_folder1 = [filename[:-len(suffix_need_remove)] + filename[-len(suffix_need_remove):].replace(suffix_need_remove, '') for filename in filenames_noext_folder1]
        if suffix_floder == 1:
            filenames_raw = list(filenames_noext_folder2).copy()
            filenames_noext_folder2 = [filename[:-len(suffix_need_remove)] + filename[-len(suffix_need_remove):].replace(suffix_need_remove, '') for filename in filenames_noext_folder2]

    for index, filename in enumerate(filenames_noext_folder1):
        if filename in filenames_noext_folder2:
            print(filename)
            if suffix_need_remove:
                if suffix_floder == 0:
                    img1 = os.path.join(folder1, filenames_raw[index] + extensions_folder1[index])
                    img2 = os.path.join(folder2, filename + extensions_folder2[filenames_noext_folder2.index(filename)])
                if suffix_floder == 1:
                    img1 = os.path.join(folder1, filename + extensions_folder1[index])
                    img2 = os.path.join(folder2, filenames_raw[filenames_noext_folder2.index(filename)] + extensions_folder2[filenames_noext_folder2.index(filename)])
            else:
                img1 = os.path.join(folder1, filename + extensions_folder1[index])
                img2 = os.path.join(folder2, filename + extensions_folder2[filenames_noext_folder2.index(filename)])
            result = merge_2_images(img1, img2, mode, erosion_para, dilate)
            cv2.imwrite(os.path.join(outdir, filename + extensions_folder1[index]), result)


def invert_image(image):
    # 将图片从BGR转为灰度图
    gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
    # 对灰度图进行反转
    inverted_image = cv2.bitwise_not(gray_image)
    # 将反转后的灰度图转换回BGR格式
    inverted_image_bgr = cv2.cvtColor(inverted_image, cv2.COLOR_GRAY2BGR)
    return inverted_image_bgr

def process_images(input_folder='./output/merged_imgs'):
    output_folder = os.path.join(os.path.dirname(input_folder), 'output_invert')
    os.makedirs(output_folder, exist_ok=True)

    # 获取输入文件夹中的所有图片文件
    image_files = [f for f in os.listdir(input_folder) if
                   f.lower().endswith(('.jpg', '.jpeg', '.png', '.gif', '.bmp', '.tiff', '.webp', 'tif'))]

    for image_file in tqdm(image_files):
        image_path = os.path.join(input_folder, image_file)
        try:
            # 使用PIL库读取图像
            with Image.open(image_path) as img:
                image = np.array(img.convert('RGB'))[:, :, ::-1].copy()
                if image is not None:
                    # 翻转图片
                    inverted_image = invert_image(image)
                    # 保存翻转后的图片到输出文件夹
                    output_path = os.path.join(output_folder, image_file)
                    cv2.imwrite(output_path, inverted_image)
                else:
                    raise ValueError(f"Failed to read image: {image_file}")
        except Exception as e:
            print(f"Error processing file {image_file}: {e}")

def process_line(img_path='./input', outdir='./output'):
    # 处理深度图像
    depth_anything(img_path, os.path.join(outdir, "depth_anything"))

    # 处理 teed_imgs
    teed_imgs(img_path, os.path.join(outdir, "teed_imgs"), [1, 7, 2])
    
    # 处理深度图的 teed_imgs
    teed_imgs(os.path.join(outdir, "depth_anything"), os.path.join(outdir, "dp_teed_imgs"), [0, 7, 2])
    
    # 合并图像
    merge_images_in_2_folder(
        os.path.join(outdir, "teed_imgs"),
        os.path.join(outdir, "dp_teed_imgs"),
        os.path.join(outdir, "merged_imgs"),
        '_depth',
        1,
        'multiply',
        [[2, 0], [2, 1]],
        [1, 0]
    )


if __name__ == '__main__':
    # Gradio 接口代码
    import gradio as gr

    def gradio_process_line(img):
        img_path = './temp_input.png'
        img.save(img_path)

        process_line(img_path, './output')

        output_image_path = './output/teed_imgs/processed_image.png'  # 更新为实际输出路径
        return Image.open(output_image_path)

    # 定义 Gradio 接口
    iface = gr.Interface(
        fn=gradio_process_line,
        inputs=gr.Image(type="pil"),
        outputs=gr.Image(type="pil"),
        title="Image Processing with Depth Anything and TEED",
        description="Upload an image to process it with depth estimation and edge detection."
    )

    # 启动 Gradio 应用
    iface.launch()