File size: 14,095 Bytes
d5e54ca
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
"""
并行执行视频评测任务

该脚本可以自动将评测任务分成多个组,并行执行,最后自动合并结果。
支持自动将多张 GPU 均匀分配给不同的任务组。

使用方法:
    # 使用 8 张 GPU,分成 8 组并行执行
    python parallel_evaluate.py \\
        --model_input_dir ./inputs \\
        --model_output_dir ./outputs \\
        --evaluate_subjects dino_consistency \\
        --group_total 8 \\
        --num_gpus 8
    
    # 使用 4 张 GPU,分成 8 组并行执行(每张 GPU 运行 2 个任务)
    python parallel_evaluate.py \\
        --model_input_dir ./inputs \\
        --model_output_dir ./outputs \\
        --evaluate_subjects dino_consistency \\
        --group_total 8 \\
        --num_gpus 4 \\
        --parallelism 4
    
    # 指定使用哪些 GPU
    python parallel_evaluate.py \\
        --model_input_dir ./inputs \\
        --model_output_dir ./outputs \\
        --evaluate_subjects dino_consistency \\
        --group_total 8 \\
        --gpu_ids 0,1,2,3,4,5,6,7
"""

import os
import sys
import pathlib
import argparse
import subprocess
import time
from concurrent.futures import ThreadPoolExecutor, as_completed
from typing import List, Tuple, Optional
import json


def run_evaluation_group(
    group_id: int,
    group_total: int,
    model_input_dir: str,
    model_output_dir: str,
    results_dir: str,
    evaluate_subjects: str,
    gpu_id: int,
    batch_size: int = 16,
    sampling: int = 0,
    model_args: str = '{}',
    filter_str: str = '',
    extra_args: List[str] = None
) -> Tuple[int, bool, str]:
    """
    运行单个评测组
    
    Args:
        group_id: 组号
        group_total: 总组数
        model_input_dir: 输入目录
        model_output_dir: 输出目录
        results_dir: 结果目录
        evaluate_subjects: 评测主题
        gpu_id: GPU ID
        batch_size: 批处理大小
        sampling: 采样帧数
        model_args: 模型参数
        filter_str: 文件名过滤
        extra_args: 额外参数列表
    
    Returns:
        (group_id, success, message): 组号、是否成功、消息
    """
    cmd = [
        'python', 'evaluate.py',
        '--model_input_dir', model_input_dir,
        '--model_output_dir', model_output_dir,
        '--results_dir', results_dir,
        '--evaluate_subjects', evaluate_subjects,
        '--device', f'cuda:{gpu_id}',
        '--batch_size', str(batch_size),
        '--sampling', str(sampling),
        '--model_args', model_args,
        '--group_id', str(group_id),
        '--group_total', str(group_total),
    ]
    
    if filter_str:
        cmd.extend(['--filter', filter_str])
    
    if extra_args:
        cmd.extend(extra_args)
    
    # 设置环境变量,指定使用的 GPU
    env = os.environ.copy()
    env['CUDA_VISIBLE_DEVICES'] = str(gpu_id)
    
    log_file = pathlib.Path(results_dir) / f'group{group_id}of{group_total}.log'
    
    try:
        with open(log_file, 'w') as f:
            f.write(f"Command: {' '.join(cmd)}\n")
            f.write(f"GPU: {gpu_id}\n")
            f.write(f"{'='*60}\n\n")
            f.flush()
            
            process = subprocess.Popen(
                cmd,
                stdout=subprocess.PIPE,
                stderr=subprocess.STDOUT,
                env=env,
                text=True,
                bufsize=1
            )
            
            # 实时写入日志
            for line in process.stdout:
                f.write(line)
                f.flush()
            
            process.wait()
            
            if process.returncode == 0:
                return group_id, True, f"Group {group_id} completed successfully"
            else:
                return group_id, False, f"Group {group_id} failed with return code {process.returncode}"
    
    except Exception as e:
        return group_id, False, f"Group {group_id} failed with exception: {str(e)}"


def merge_results(results_dir: str, subjects: str, group_total: int, output_file: str = None) -> bool:
    """
    调用 merge_results.py 合并结果
    
    Args:
        results_dir: 结果目录
        subjects: 评测主题
        group_total: 总组数
        output_file: 输出文件路径(可选)
    
    Returns:
        是否成功
    """
    cmd = [
        'python', 'merge_results.py',
        '--results_dir', results_dir,
        '--subjects', subjects,
        '--group_total', str(group_total)
    ]
    
    if output_file:
        cmd.extend(['--output', output_file])
    
    try:
        result = subprocess.run(cmd, check=True, capture_output=True, text=True)
        print(result.stdout)
        return True
    except subprocess.CalledProcessError as e:
        print(f"Merge failed: {e}")
        print(e.stderr)
        return False


def main():
    parser = argparse.ArgumentParser(
        description="Parallel evaluation of video generation models with automatic GPU allocation.",
        formatter_class=argparse.RawDescriptionHelpFormatter,
        epilog="""
Examples:
  # Use 8 GPUs for 8 groups (one group per GPU)
  python parallel_evaluate.py --model_input_dir ./inputs --model_output_dir ./outputs \\
      --evaluate_subjects dino_consistency --group_total 8 --num_gpus 8
  
  # Use 4 GPUs for 8 groups with parallelism of 4 (2 groups per GPU)
  python parallel_evaluate.py --model_input_dir ./inputs --model_output_dir ./outputs \\
      --evaluate_subjects dino_consistency --group_total 8 --num_gpus 4 --parallelism 4
  
  # Specify which GPUs to use
  python parallel_evaluate.py --model_input_dir ./inputs --model_output_dir ./outputs \\
      --evaluate_subjects dino_consistency --group_total 8 --gpu_ids 0,2,4,6
        """
    )
    
    # 必需参数
    parser.add_argument('--model_input_dir', type=str, required=True,
                        help='Path to the input directory of the evaluated model.')
    parser.add_argument('--model_output_dir', type=str, required=True,
                        help='Path to the output directory of the evaluated model.')
    parser.add_argument('--evaluate_subjects', type=str, required=True,
                        help='Comma-separated list of evaluation subjects.')
    
    # 分组参数
    parser.add_argument('--group_total', type=int, required=True,
                        help='Total number of groups to split the evaluation into.')
    
    # GPU 配置
    gpu_group = parser.add_mutually_exclusive_group()
    gpu_group.add_argument('--num_gpus', type=int, default=8,
                          help='Number of GPUs to use (default: 8). GPUs will be numbered 0 to num_gpus-1.')
    gpu_group.add_argument('--gpu_ids', type=str,
                          help='Comma-separated list of GPU IDs to use, e.g., "0,1,2,3,4,5,6,7".')
    
    # 并行度
    parser.add_argument('--parallelism', type=int, default=None,
                        help='Maximum number of parallel tasks. Default is equal to number of GPUs.')
    
    # 评测参数
    parser.add_argument('--results_dir', type=str, default='./evaluation_results',
                        help='Path to the directory where evaluation results will be saved.')
    parser.add_argument('--batch_size', type=int, default=16,
                        help='Batch size for evaluation (default: 16).')
    parser.add_argument('--sampling', type=int, default=0,
                        help='Number of frames to sample (default: 0, use all frames).')
    parser.add_argument('--model_args', type=str, default='{}',
                        help='Additional model arguments in JSON format.')
    parser.add_argument('--filter', type=str, default='',
                        help='Filter to only evaluate videos whose filenames contain this string.')
    
    # 合并结果
    parser.add_argument('--output', type=str, default=None,
                        help='Output file path for merged results.')
    parser.add_argument('--skip_merge', action='store_true',
                        help='Skip merging results after all groups complete.')
    
    # 其他选项
    parser.add_argument('--dry_run', action='store_true',
                        help='Print commands without executing them.')
    
    args = parser.parse_args()
    
    # 解析 GPU IDs
    if args.gpu_ids:
        gpu_ids = [int(x.strip()) for x in args.gpu_ids.split(',')]
    else:
        gpu_ids = list(range(args.num_gpus))
    
    num_gpus = len(gpu_ids)
    
    # 设置并行度
    if args.parallelism is None:
        parallelism = num_gpus
    else:
        parallelism = args.parallelism
    
    # 验证参数
    if args.group_total <= 0:
        print("Error: --group_total must be greater than 0")
        sys.exit(1)
    
    if parallelism <= 0:
        print("Error: --parallelism must be greater than 0")
        sys.exit(1)
    
    # 创建结果目录
    results_dir = pathlib.Path(args.results_dir)
    results_dir.mkdir(parents=True, exist_ok=True)
    
    print(f"{'='*60}")
    print(f"Parallel Evaluation Configuration:")
    print(f"{'='*60}")
    print(f"Model Input Dir:     {args.model_input_dir}")
    print(f"Model Output Dir:    {args.model_output_dir}")
    print(f"Results Dir:         {args.results_dir}")
    print(f"Evaluate Subjects:   {args.evaluate_subjects}")
    print(f"Total Groups:        {args.group_total}")
    print(f"Available GPUs:      {gpu_ids}")
    print(f"Number of GPUs:      {num_gpus}")
    print(f"Parallelism:         {parallelism}")
    print(f"Batch Size:          {args.batch_size}")
    print(f"Sampling:            {args.sampling}")
    print(f"{'='*60}\n")
    
    if args.dry_run:
        print("DRY RUN MODE - Commands that would be executed:")
        for group_id in range(args.group_total):
            gpu_id = gpu_ids[group_id % num_gpus]
            print(f"\nGroup {group_id} (GPU {gpu_id}):")
            cmd = [
                'python', 'evaluate.py',
                '--model_input_dir', args.model_input_dir,
                '--model_output_dir', args.model_output_dir,
                '--results_dir', args.results_dir,
                '--evaluate_subjects', args.evaluate_subjects,
                '--device', f'cuda:{gpu_id}',
                '--batch_size', str(args.batch_size),
                '--sampling', str(args.sampling),
                '--model_args', args.model_args,
                '--group_id', str(group_id),
                '--group_total', str(args.group_total),
            ]
            if args.filter:
                cmd.extend(['--filter', args.filter])
            print('  ' + ' '.join(cmd))
        print("\nDry run complete. No tasks were executed.")
        return
    
    # 执行并行评测
    start_time = time.time()
    completed_groups = []
    failed_groups = []
    
    with ThreadPoolExecutor(max_workers=parallelism) as executor:
        # 提交所有任务
        future_to_group = {}
        for group_id in range(args.group_total):
            # 循环分配 GPU
            gpu_id = gpu_ids[group_id % num_gpus]
            
            future = executor.submit(
                run_evaluation_group,
                group_id=group_id,
                group_total=args.group_total,
                model_input_dir=args.model_input_dir,
                model_output_dir=args.model_output_dir,
                results_dir=args.results_dir,
                evaluate_subjects=args.evaluate_subjects,
                gpu_id=gpu_id,
                batch_size=args.batch_size,
                sampling=args.sampling,
                model_args=args.model_args,
                filter_str=args.filter
            )
            future_to_group[future] = group_id
            print(f"Submitted group {group_id} to GPU {gpu_id}")
        
        print(f"\nAll {args.group_total} groups submitted. Waiting for completion...\n")
        
        # 收集结果
        for future in as_completed(future_to_group):
            group_id, success, message = future.result()
            
            if success:
                completed_groups.append(group_id)
                print(f"✓ {message}")
            else:
                failed_groups.append(group_id)
                print(f"✗ {message}")
            
            print(f"  Progress: {len(completed_groups) + len(failed_groups)}/{args.group_total} "
                  f"(Success: {len(completed_groups)}, Failed: {len(failed_groups)})\n")
    
    elapsed_time = time.time() - start_time
    
    # 打印总结
    print(f"\n{'='*60}")
    print(f"Evaluation Summary:")
    print(f"{'='*60}")
    print(f"Total groups:        {args.group_total}")
    print(f"Completed:           {len(completed_groups)}")
    print(f"Failed:              {len(failed_groups)}")
    print(f"Time elapsed:        {elapsed_time:.2f} seconds ({elapsed_time/60:.2f} minutes)")
    
    if failed_groups:
        print(f"\nFailed groups: {sorted(failed_groups)}")
        print("Check log files for details.")
    
    # 合并结果
    if not args.skip_merge and len(completed_groups) > 0:
        print(f"\n{'='*60}")
        print("Merging results...")
        print(f"{'='*60}\n")
        
        merge_success = merge_results(
            results_dir=args.results_dir,
            subjects=args.evaluate_subjects,
            group_total=args.group_total,
            output_file=args.output
        )
        
        if merge_success:
            print("\n✓ All results merged successfully!")
        else:
            print("\n✗ Failed to merge results. You can try merging manually with merge_results.py")
            sys.exit(1)
    elif args.skip_merge:
        print("\nSkipping merge (--skip_merge specified)")
    else:
        print("\nNo results to merge (all groups failed)")
        sys.exit(1)
    
    if failed_groups:
        print(f"\n⚠ Warning: {len(failed_groups)} group(s) failed. Check log files.")
        sys.exit(1)
    else:
        print(f"\n{'='*60}")
        print("✓ All tasks completed successfully!")
        print(f"{'='*60}")


if __name__ == "__main__":
    main()