Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 144, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 134, in main run_eval(eval_args, mode="compute_metrics", verbose=True) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 70, in run_eval print("Model: ", args.model_path, "device: ", model.device) AttributeError: 'NoneType' object has no attribute 'device' Model: /fsx_0/user/imzyc/proact_exps/20240822-L4096-I5-ep4-NOSEP-nr0.1-klgmix-1s-lora-bs384-debug {'assembly101/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}, 'ego4d/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.05}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}, 'egoexolearn/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}, 'epickitchens/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}, 'holoassist/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}, 'wtag/dialog_val_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.1}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}, {'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}} Evaluation datasets: * ego4d/dialog_val | num samples: 96 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.05 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.05-maxlen_4k Metrics: missing_rate: 0.8143 redundant_rate: 0.0418 semantic_score: 0.7024 jaccard_index: 0.1469 precision: 0.7642 recall: 0.1481 F1: 0.2481 num_matched: 713.0000 num_mismatched: 181.0000 num_missed: 3920.0000 num_redundant: 39.0000 Bleu_1: 0.4112 Bleu_1_w: 0.0604 Bleu_2: 0.3002 Bleu_2_w: 0.0441 Bleu_3: 0.2331 Bleu_3_w: 0.0342 Bleu_4: 0.1877 Bleu_4_w: 0.0276 CIDEr: 1.1146 CIDEr_w: 0.1638 METEOR: 0.2052 METEOR_w: 0.0302 Updating eval setup: not_talk_threshold: 0.05 -> 0.1 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.7900 redundant_rate: 0.1124 semantic_score: 0.7032 jaccard_index: 0.1566 precision: 0.6795 recall: 0.1608 F1: 0.2600 num_matched: 774.0000 num_mismatched: 237.0000 num_missed: 3803.0000 num_redundant: 128.0000 Bleu_1: 0.4120 Bleu_1_w: 0.0645 Bleu_2: 0.2990 Bleu_2_w: 0.0468 Bleu_3: 0.2308 Bleu_3_w: 0.0362 Bleu_4: 0.1849 Bleu_4_w: 0.0290 CIDEr: 1.1293 CIDEr_w: 0.1769 METEOR: 0.2049 METEOR_w: 0.0321 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.7179 redundant_rate: 0.2687 semantic_score: 0.6990 jaccard_index: 0.1850 precision: 0.5293 recall: 0.2042 F1: 0.2947 num_matched: 983.0000 num_mismatched: 375.0000 num_missed: 3456.0000 num_redundant: 499.0000 Bleu_1: 0.4005 Bleu_1_w: 0.0741 Bleu_2: 0.2853 Bleu_2_w: 0.0528 Bleu_3: 0.2176 Bleu_3_w: 0.0403 Bleu_4: 0.1735 Bleu_4_w: 0.0321 CIDEr: 0.9623 CIDEr_w: 0.1780 METEOR: 0.1921 METEOR_w: 0.0355 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.6342 redundant_rate: 0.4073 semantic_score: 0.6859 jaccard_index: 0.2117 precision: 0.4291 recall: 0.2649 F1: 0.3276 num_matched: 1275.0000 num_mismatched: 486.0000 num_missed: 3053.0000 num_redundant: 1210.0000 Bleu_1: 0.3777 Bleu_1_w: 0.0799 Bleu_2: 0.2595 Bleu_2_w: 0.0549 Bleu_3: 0.1918 Bleu_3_w: 0.0406 Bleu_4: 0.1491 Bleu_4_w: 0.0316 CIDEr: 0.7785 CIDEr_w: 0.1648 METEOR: 0.1803 METEOR_w: 0.0382 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.5374 redundant_rate: 0.5324 semantic_score: 0.6808 jaccard_index: 0.2079 precision: 0.3208 recall: 0.3174 F1: 0.3191 num_matched: 1528.0000 num_mismatched: 699.0000 num_missed: 2587.0000 num_redundant: 2536.0000 Bleu_1: 0.3763 Bleu_1_w: 0.0782 Bleu_2: 0.2556 Bleu_2_w: 0.0531 Bleu_3: 0.1874 Bleu_3_w: 0.0390 Bleu_4: 0.1447 Bleu_4_w: 0.0301 CIDEr: 0.7667 CIDEr_w: 0.1594 METEOR: 0.1752 METEOR_w: 0.0364 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: ego4d-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.4053 redundant_rate: 0.7022 semantic_score: 0.6725 jaccard_index: 0.1596 precision: 0.1920 recall: 0.3835 F1: 0.2559 num_matched: 1846.0000 num_mismatched: 1017.0000 num_missed: 1951.0000 num_redundant: 6750.0000 Bleu_1: 0.3556 Bleu_1_w: 0.0568 Bleu_2: 0.2372 Bleu_2_w: 0.0379 Bleu_3: 0.1705 Bleu_3_w: 0.0272 Bleu_4: 0.1289 Bleu_4_w: 0.0206 CIDEr: 0.6627 CIDEr_w: 0.1058 METEOR: 0.1699 METEOR_w: 0.0271 Evaluation datasets: * holoassist/dialog_val | num samples: 291 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.1 Evalulation: holoassist-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.8169 redundant_rate: 0.0096 semantic_score: 0.6931 jaccard_index: 0.1324 precision: 0.7175 recall: 0.1326 F1: 0.2239 num_matched: 2024.0000 num_mismatched: 770.0000 num_missed: 12467.0000 num_redundant: 27.0000 Bleu_1: 0.4319 Bleu_1_w: 0.0572 Bleu_2: 0.3132 Bleu_2_w: 0.0415 Bleu_3: 0.2389 Bleu_3_w: 0.0316 Bleu_4: 0.1871 Bleu_4_w: 0.0248 CIDEr: 1.1122 CIDEr_w: 0.1472 METEOR: 0.2072 METEOR_w: 0.0274 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: holoassist-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.6569 redundant_rate: 0.0176 semantic_score: 0.6920 jaccard_index: 0.2380 precision: 0.6856 recall: 0.2394 F1: 0.3549 num_matched: 3654.0000 num_mismatched: 1582.0000 num_missed: 10025.0000 num_redundant: 94.0000 Bleu_1: 0.4265 Bleu_1_w: 0.1015 Bleu_2: 0.3066 Bleu_2_w: 0.0730 Bleu_3: 0.2318 Bleu_3_w: 0.0552 Bleu_4: 0.1807 Bleu_4_w: 0.0430 CIDEr: 1.0739 CIDEr_w: 0.2556 METEOR: 0.2021 METEOR_w: 0.0481 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: holoassist-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.5975 redundant_rate: 0.0346 semantic_score: 0.6902 jaccard_index: 0.2730 precision: 0.6643 recall: 0.2770 F1: 0.3910 num_matched: 4227.0000 num_mismatched: 1916.0000 num_missed: 9118.0000 num_redundant: 220.0000 Bleu_1: 0.4227 Bleu_1_w: 0.1154 Bleu_2: 0.3021 Bleu_2_w: 0.0825 Bleu_3: 0.2281 Bleu_3_w: 0.0623 Bleu_4: 0.1775 Bleu_4_w: 0.0485 CIDEr: 1.0430 CIDEr_w: 0.2848 METEOR: 0.1995 METEOR_w: 0.0545 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: holoassist-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.4488 redundant_rate: 0.3363 semantic_score: 0.6851 jaccard_index: 0.2771 precision: 0.4268 recall: 0.3544 F1: 0.3873 num_matched: 5409.0000 num_mismatched: 3003.0000 num_missed: 6849.0000 num_redundant: 4262.0000 Bleu_1: 0.4084 Bleu_1_w: 0.1132 Bleu_2: 0.2863 Bleu_2_w: 0.0793 Bleu_3: 0.2127 Bleu_3_w: 0.0589 Bleu_4: 0.1632 Bleu_4_w: 0.0452 CIDEr: 0.9756 CIDEr_w: 0.2703 METEOR: 0.1921 METEOR_w: 0.0532 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: holoassist-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.2812 redundant_rate: 0.6232 semantic_score: 0.6782 jaccard_index: 0.2047 precision: 0.2349 recall: 0.4481 F1: 0.3082 num_matched: 6838.0000 num_mismatched: 4131.0000 num_missed: 4292.0000 num_redundant: 18140.0000 Bleu_1: 0.3886 Bleu_1_w: 0.0795 Bleu_2: 0.2671 Bleu_2_w: 0.0547 Bleu_3: 0.1963 Bleu_3_w: 0.0402 Bleu_4: 0.1491 Bleu_4_w: 0.0305 CIDEr: 0.8726 CIDEr_w: 0.1786 METEOR: 0.1812 METEOR_w: 0.0371 Evaluation datasets: * epickitchens/dialog_val | num samples: 150 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.1 Evalulation: epickitchens-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.7505 redundant_rate: 0.0771 semantic_score: 0.6804 jaccard_index: 0.1637 precision: 0.6182 recall: 0.1671 F1: 0.2631 num_matched: 1075.0000 num_mismatched: 530.0000 num_missed: 4827.0000 num_redundant: 134.0000 Bleu_1: 0.4001 Bleu_1_w: 0.0655 Bleu_2: 0.2872 Bleu_2_w: 0.0470 Bleu_3: 0.2182 Bleu_3_w: 0.0357 Bleu_4: 0.1706 Bleu_4_w: 0.0279 CIDEr: 1.1686 CIDEr_w: 0.1913 METEOR: 0.2001 METEOR_w: 0.0328 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: epickitchens-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.6191 redundant_rate: 0.1988 semantic_score: 0.6640 jaccard_index: 0.2061 precision: 0.4745 recall: 0.2256 F1: 0.3058 num_matched: 1451.0000 num_mismatched: 999.0000 num_missed: 3982.0000 num_redundant: 608.0000 Bleu_1: 0.3887 Bleu_1_w: 0.0801 Bleu_2: 0.2711 Bleu_2_w: 0.0559 Bleu_3: 0.2005 Bleu_3_w: 0.0413 Bleu_4: 0.1546 Bleu_4_w: 0.0319 CIDEr: 1.0446 CIDEr_w: 0.2153 METEOR: 0.1885 METEOR_w: 0.0389 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: epickitchens-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.4639 redundant_rate: 0.3537 semantic_score: 0.6534 jaccard_index: 0.2261 precision: 0.3526 recall: 0.2924 F1: 0.3197 num_matched: 1881.0000 num_mismatched: 1567.0000 num_missed: 2984.0000 num_redundant: 1887.0000 Bleu_1: 0.3547 Bleu_1_w: 0.0802 Bleu_2: 0.2373 Bleu_2_w: 0.0537 Bleu_3: 0.1686 Bleu_3_w: 0.0381 Bleu_4: 0.1262 Bleu_4_w: 0.0285 CIDEr: 0.8348 CIDEr_w: 0.1888 METEOR: 0.1764 METEOR_w: 0.0399 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: epickitchens-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.3136 redundant_rate: 0.5891 semantic_score: 0.6491 jaccard_index: 0.1801 precision: 0.2138 recall: 0.3573 F1: 0.2676 num_matched: 2298.0000 num_mismatched: 2117.0000 num_missed: 2017.0000 num_redundant: 6331.0000 Bleu_1: 0.3525 Bleu_1_w: 0.0635 Bleu_2: 0.2293 Bleu_2_w: 0.0413 Bleu_3: 0.1598 Bleu_3_w: 0.0288 Bleu_4: 0.1185 Bleu_4_w: 0.0213 CIDEr: 0.7864 CIDEr_w: 0.1416 METEOR: 0.1703 METEOR_w: 0.0307 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: epickitchens-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.2040 redundant_rate: 0.7203 semantic_score: 0.6452 jaccard_index: 0.1408 precision: 0.1509 recall: 0.4296 F1: 0.2234 num_matched: 2763.0000 num_mismatched: 2357.0000 num_missed: 1312.0000 num_redundant: 13187.0000 Bleu_1: 0.3569 Bleu_1_w: 0.0503 Bleu_2: 0.2305 Bleu_2_w: 0.0325 Bleu_3: 0.1585 Bleu_3_w: 0.0223 Bleu_4: 0.1160 Bleu_4_w: 0.0163 CIDEr: 0.7808 CIDEr_w: 0.1100 METEOR: 0.1699 METEOR_w: 0.0239 Evaluation datasets: * egoexolearn/dialog_val | num samples: 123 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.1 Evalulation: egoexolearn-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.8310 redundant_rate: 0.0334 semantic_score: 0.6980 jaccard_index: 0.1260 precision: 0.7252 recall: 0.1268 F1: 0.2158 num_matched: 1520.0000 num_mismatched: 506.0000 num_missed: 9965.0000 num_redundant: 70.0000 Bleu_1: 0.4299 Bleu_1_w: 0.0542 Bleu_2: 0.3105 Bleu_2_w: 0.0391 Bleu_3: 0.2375 Bleu_3_w: 0.0299 Bleu_4: 0.1875 Bleu_4_w: 0.0236 CIDEr: 1.1086 CIDEr_w: 0.1397 METEOR: 0.2051 METEOR_w: 0.0258 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: egoexolearn-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.8105 redundant_rate: 0.1288 semantic_score: 0.6944 jaccard_index: 0.1330 precision: 0.6288 recall: 0.1368 F1: 0.2247 num_matched: 1640.0000 num_mismatched: 632.0000 num_missed: 9719.0000 num_redundant: 336.0000 Bleu_1: 0.4263 Bleu_1_w: 0.0567 Bleu_2: 0.3069 Bleu_2_w: 0.0408 Bleu_3: 0.2339 Bleu_3_w: 0.0311 Bleu_4: 0.1840 Bleu_4_w: 0.0245 CIDEr: 1.0771 CIDEr_w: 0.1433 METEOR: 0.2020 METEOR_w: 0.0269 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: egoexolearn-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.7804 redundant_rate: 0.2401 semantic_score: 0.6867 jaccard_index: 0.1436 precision: 0.5313 recall: 0.1535 F1: 0.2382 num_matched: 1841.0000 num_mismatched: 792.0000 num_missed: 9358.0000 num_redundant: 832.0000 Bleu_1: 0.4164 Bleu_1_w: 0.0598 Bleu_2: 0.2942 Bleu_2_w: 0.0422 Bleu_3: 0.2212 Bleu_3_w: 0.0318 Bleu_4: 0.1721 Bleu_4_w: 0.0247 CIDEr: 1.0102 CIDEr_w: 0.1450 METEOR: 0.1951 METEOR_w: 0.0280 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: egoexolearn-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.7114 redundant_rate: 0.3865 semantic_score: 0.6745 jaccard_index: 0.1580 precision: 0.3969 recall: 0.1867 F1: 0.2540 num_matched: 2239.0000 num_mismatched: 1222.0000 num_missed: 8530.0000 num_redundant: 2180.0000 Bleu_1: 0.4039 Bleu_1_w: 0.0638 Bleu_2: 0.2814 Bleu_2_w: 0.0445 Bleu_3: 0.2088 Bleu_3_w: 0.0330 Bleu_4: 0.1606 Bleu_4_w: 0.0254 CIDEr: 0.9108 CIDEr_w: 0.1439 METEOR: 0.1870 METEOR_w: 0.0296 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: egoexolearn-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.5728 redundant_rate: 0.5894 semantic_score: 0.6633 jaccard_index: 0.1569 precision: 0.2432 recall: 0.2531 F1: 0.2481 num_matched: 3035.0000 num_mismatched: 2088.0000 num_missed: 6868.0000 num_redundant: 7355.0000 Bleu_1: 0.3851 Bleu_1_w: 0.0604 Bleu_2: 0.2611 Bleu_2_w: 0.0410 Bleu_3: 0.1877 Bleu_3_w: 0.0294 Bleu_4: 0.1406 Bleu_4_w: 0.0221 CIDEr: 0.7626 CIDEr_w: 0.1196 METEOR: 0.1741 METEOR_w: 0.0273 Evaluation datasets: * wtag/dialog_val | num samples: 21 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.1 Evalulation: wtag-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.5732 redundant_rate: 0.1089 semantic_score: 0.6824 jaccard_index: 0.2099 precision: 0.4611 recall: 0.2209 F1: 0.2987 num_matched: 237.0000 num_mismatched: 221.0000 num_missed: 615.0000 num_redundant: 56.0000 Bleu_1: 0.3633 Bleu_1_w: 0.0763 Bleu_2: 0.2567 Bleu_2_w: 0.0539 Bleu_3: 0.1885 Bleu_3_w: 0.0396 Bleu_4: 0.1448 Bleu_4_w: 0.0304 CIDEr: 0.9644 CIDEr_w: 0.2025 METEOR: 0.2138 METEOR_w: 0.0449 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: wtag-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.5005 redundant_rate: 0.1612 semantic_score: 0.6793 jaccard_index: 0.2109 precision: 0.3881 recall: 0.2311 F1: 0.2897 num_matched: 248.0000 num_mismatched: 288.0000 num_missed: 537.0000 num_redundant: 103.0000 Bleu_1: 0.3624 Bleu_1_w: 0.0764 Bleu_2: 0.2559 Bleu_2_w: 0.0540 Bleu_3: 0.1871 Bleu_3_w: 0.0395 Bleu_4: 0.1435 Bleu_4_w: 0.0303 CIDEr: 0.9300 CIDEr_w: 0.1961 METEOR: 0.2112 METEOR_w: 0.0445 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: wtag-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.4352 redundant_rate: 0.1844 semantic_score: 0.6717 jaccard_index: 0.2099 precision: 0.3419 recall: 0.2367 F1: 0.2797 num_matched: 254.0000 num_mismatched: 352.0000 num_missed: 467.0000 num_redundant: 137.0000 Bleu_1: 0.3506 Bleu_1_w: 0.0736 Bleu_2: 0.2455 Bleu_2_w: 0.0515 Bleu_3: 0.1793 Bleu_3_w: 0.0376 Bleu_4: 0.1362 Bleu_4_w: 0.0286 CIDEr: 0.8458 CIDEr_w: 0.1775 METEOR: 0.2037 METEOR_w: 0.0428 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: wtag-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.4101 redundant_rate: 0.2383 semantic_score: 0.6682 jaccard_index: 0.2242 precision: 0.3430 recall: 0.2656 F1: 0.2994 num_matched: 285.0000 num_mismatched: 348.0000 num_missed: 440.0000 num_redundant: 198.0000 Bleu_1: 0.3571 Bleu_1_w: 0.0801 Bleu_2: 0.2481 Bleu_2_w: 0.0556 Bleu_3: 0.1788 Bleu_3_w: 0.0401 Bleu_4: 0.1315 Bleu_4_w: 0.0295 CIDEr: 0.8673 CIDEr_w: 0.1945 METEOR: 0.2046 METEOR_w: 0.0459 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: wtag-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.3849 redundant_rate: 0.3038 semantic_score: 0.6623 jaccard_index: 0.2035 precision: 0.2922 recall: 0.2582 F1: 0.2741 num_matched: 277.0000 num_mismatched: 383.0000 num_missed: 413.0000 num_redundant: 288.0000 Bleu_1: 0.3455 Bleu_1_w: 0.0703 Bleu_2: 0.2390 Bleu_2_w: 0.0486 Bleu_3: 0.1720 Bleu_3_w: 0.0350 Bleu_4: 0.1264 Bleu_4_w: 0.0257 CIDEr: 0.7777 CIDEr_w: 0.1583 METEOR: 0.1970 METEOR_w: 0.0401 Evaluation datasets: * assembly101/dialog_val | num samples: 336 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.1 Evalulation: assembly101-dialog_val_L0_I5/stream/notalk0.1-maxlen_4k Metrics: missing_rate: 0.7505 redundant_rate: 0.0598 semantic_score: 0.6999 jaccard_index: 0.1639 precision: 0.6275 recall: 0.1665 F1: 0.2632 num_matched: 1385.0000 num_mismatched: 690.0000 num_missed: 6243.0000 num_redundant: 132.0000 Bleu_1: 0.4494 Bleu_1_w: 0.0737 Bleu_2: 0.3420 Bleu_2_w: 0.0560 Bleu_3: 0.2687 Bleu_3_w: 0.0440 Bleu_4: 0.2174 Bleu_4_w: 0.0356 CIDEr: 1.2016 CIDEr_w: 0.1969 METEOR: 0.2251 METEOR_w: 0.0369 Updating eval setup: not_talk_threshold: 0.1 -> 0.2 Evalulation: assembly101-dialog_val_L0_I5/stream/notalk0.2-maxlen_4k Metrics: missing_rate: 0.6718 redundant_rate: 0.1407 semantic_score: 0.6904 jaccard_index: 0.1954 precision: 0.5392 recall: 0.2059 F1: 0.2980 num_matched: 1713.0000 num_mismatched: 1017.0000 num_missed: 5588.0000 num_redundant: 447.0000 Bleu_1: 0.4380 Bleu_1_w: 0.0856 Bleu_2: 0.3286 Bleu_2_w: 0.0642 Bleu_3: 0.2547 Bleu_3_w: 0.0498 Bleu_4: 0.2038 Bleu_4_w: 0.0398 CIDEr: 1.1287 CIDEr_w: 0.2206 METEOR: 0.2136 METEOR_w: 0.0417 Updating eval setup: not_talk_threshold: 0.2 -> 0.3 Evalulation: assembly101-dialog_val_L0_I5/stream/notalk0.3-maxlen_4k Metrics: missing_rate: 0.5380 redundant_rate: 0.2288 semantic_score: 0.6756 jaccard_index: 0.2341 precision: 0.4443 recall: 0.2662 F1: 0.3329 num_matched: 2214.0000 num_mismatched: 1629.0000 num_missed: 4475.0000 num_redundant: 1140.0000 Bleu_1: 0.4198 Bleu_1_w: 0.0983 Bleu_2: 0.3065 Bleu_2_w: 0.0718 Bleu_3: 0.2322 Bleu_3_w: 0.0544 Bleu_4: 0.1824 Bleu_4_w: 0.0427 CIDEr: 0.9634 CIDEr_w: 0.2255 METEOR: 0.2017 METEOR_w: 0.0472 Updating eval setup: not_talk_threshold: 0.3 -> 0.4 Evalulation: assembly101-dialog_val_L0_I5/stream/notalk0.4-maxlen_4k Metrics: missing_rate: 0.4035 redundant_rate: 0.3546 semantic_score: 0.6672 jaccard_index: 0.2383 precision: 0.3424 recall: 0.3164 F1: 0.3289 num_matched: 2632.0000 num_mismatched: 2330.0000 num_missed: 3356.0000 num_redundant: 2726.0000 Bleu_1: 0.4067 Bleu_1_w: 0.0969 Bleu_2: 0.2925 Bleu_2_w: 0.0697 Bleu_3: 0.2197 Bleu_3_w: 0.0523 Bleu_4: 0.1718 Bleu_4_w: 0.0409 CIDEr: 0.8919 CIDEr_w: 0.2126 METEOR: 0.1935 METEOR_w: 0.0461 Updating eval setup: not_talk_threshold: 0.4 -> 0.5 Evalulation: assembly101-dialog_val_L0_I5/stream/notalk0.5-maxlen_4k Metrics: missing_rate: 0.2394 redundant_rate: 0.5560 semantic_score: 0.6602 jaccard_index: 0.2055 precision: 0.2342 recall: 0.4012 F1: 0.2957 num_matched: 3337.0000 num_mismatched: 2990.0000 num_missed: 1991.0000 num_redundant: 7922.0000 Bleu_1: 0.3875 Bleu_1_w: 0.0796 Bleu_2: 0.2718 Bleu_2_w: 0.0558 Bleu_3: 0.2006 Bleu_3_w: 0.0412 Bleu_4: 0.1551 Bleu_4_w: 0.0319 CIDEr: 0.7711 CIDEr_w: 0.1584 METEOR: 0.1843 METEOR_w: 0.0379 All Finished! Time: 117.22 minutes Model: /fsx_0/user/imzyc/proact_exps/20240822-L4096-I5-ep4-NOSEP-nr0.1-klgmix-1s-lora-bs384-debug Runs: ego4d/dialog_val_L0_I5|stream|4k|0.05|summarize_and_drop ego4d/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop holoassist/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop epickitchens/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop egoexolearn/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop wtag/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop assembly101/dialog_val_L0_I5|stream|4k|0.1|summarize_and_drop ego4d/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop holoassist/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop epickitchens/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop egoexolearn/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop wtag/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop assembly101/dialog_val_L0_I5|stream|4k|0.2|summarize_and_drop ego4d/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop holoassist/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop epickitchens/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop egoexolearn/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop wtag/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop assembly101/dialog_val_L0_I5|stream|4k|0.3|summarize_and_drop ego4d/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop holoassist/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop epickitchens/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop egoexolearn/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop wtag/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop assembly101/dialog_val_L0_I5|stream|4k|0.4|summarize_and_drop ego4d/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop holoassist/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop epickitchens/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop egoexolearn/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop wtag/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop assembly101/dialog_val_L0_I5|stream|4k|0.5|summarize_and_drop sacct: error: _open_persist_conn: failed to open persistent connection to host:slurmdbd:6819: Connection refused sacct: error: Sending PersistInit msg: Connection refused sacct: error: Problem talking to the database: Connection refused submitit WARNING (2024-08-22 15:27:43,999) - Call #9 - Bypassing sacct error Command '['sacct', '-o', 'JobID,State,NodeList', '--parsable2', '-j', '14291']' returned non-zero exit status 1., status may be inaccurate. submitit WARNING (2024-08-22 15:27:43,999) - Call #9 - Bypassing sacct error Command '['sacct', '-o', 'JobID,State,NodeList', '--parsable2', '-j', '14291']' returned non-zero exit status 1., status may be inaccurate. Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 144, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 133, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14293 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14293/14293_0_result.pkl has not produced any output (state: CANCELLED by 649731) Error stream produced: ---------------------------------------- Loading checkpoint shards: 0%| | 0/4 [00:00 main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 133, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14350 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14350/14350_0_result.pkl has not produced any output (state: CANCELLED by 636977) Error stream produced: ---------------------------------------- Loading checkpoint shards: 0%| | 0/4 [00:00 main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 153, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14390 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14390/14390_0_result.pkl has not produced any output (state: NODE_FAIL) No output/error stream produced ! Check: /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14390/14390_0_log.out Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 164, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 153, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14391 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14391/14391_0_result.pkl has not produced any output (state: NODE_FAIL) No output/error stream produced ! Check: /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14391/14391_0_log.out Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 164, in print(f"Runs:\n{runs}") File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 153, in main ) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14393 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14393/14393_0_result.pkl has not produced any output (state: NODE_FAIL) Error stream produced: ---------------------------------------- slurmstepd: error: *** JOB 14393 ON h100-st-p548xlarge-129 CANCELLED AT 2024-08-22T20:08:33 DUE TO NODE FAILURE, SEE SLURMCTLD LOG FOR DETAILS *** sbatch: error: Batch job submission failed: Invalid node name specified subprocess.CalledProcessError: Command '['sbatch', '/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/submission_file_d6a19adc4ffa4a628e5fafb456cb8832.sh']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 170, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 158, in main job = executor.submit(run_eval, eval_args, "slurm_inference", verbose=True) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 734, in submit job = self._internal_process_submissions([ds])[0] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/auto/auto.py", line 218, in _internal_process_submissions return self._executor._internal_process_submissions(delayed_submissions) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/slurm/slurm.py", line 317, in _internal_process_submissions return super()._internal_process_submissions(delayed_submissions) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 893, in _internal_process_submissions job = self._submit_command(self._submitit_command_str) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 934, in _submit_command output = utils.CommandFunction(command_list, verbose=False)() # explicit errors File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/utils.py", line 354, in __call__ raise FailedJobError(stderr) from subprocess_error submitit.core.utils.FailedJobError: sbatch: error: Batch job submission failed: Invalid node name specified sbatch: error: Batch job submission failed: Invalid node name specified subprocess.CalledProcessError: Command '['sbatch', '/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/submission_file_b1632bebd8ee497f9f186b483d3918b7.sh']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 170, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 158, in main job = executor.submit(run_eval, eval_args, "slurm_inference", verbose=True) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 734, in submit job = self._internal_process_submissions([ds])[0] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/auto/auto.py", line 218, in _internal_process_submissions return self._executor._internal_process_submissions(delayed_submissions) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/slurm/slurm.py", line 317, in _internal_process_submissions return super()._internal_process_submissions(delayed_submissions) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 893, in _internal_process_submissions job = self._submit_command(self._submitit_command_str) File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 934, in _submit_command output = utils.CommandFunction(command_list, verbose=False)() # explicit errors File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/utils.py", line 354, in __call__ raise FailedJobError(stderr) from subprocess_error submitit.core.utils.FailedJobError: sbatch: error: Batch job submission failed: Invalid node name specified Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 170, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 159, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14416 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14416/14416_0_result.pkl has not produced any output (state: NODE_FAIL) Error stream produced: ---------------------------------------- slurmstepd: error: *** JOB 14416 ON h100-st-p548xlarge-2 CANCELLED AT 2024-08-22T21:17:45 DUE TO NODE FAILURE, SEE SLURMCTLD LOG FOR DETAILS *** Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 170, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 159, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14419 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14419/14419_0_result.pkl has not produced any output (state: NODE_FAIL) No output/error stream produced ! Check: /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14419/14419_0_log.out Traceback (most recent call last): File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 167, in main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 156, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14650 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14650/14650_0_result.pkl has not produced any output (state: NODE_FAIL) Error stream produced: ---------------------------------------- slurmstepd: error: *** JOB 14650 ON h100-st-p548xlarge-2 CANCELLED AT 2024-08-23T04:38:10 DUE TO NODE FAILURE, SEE SLURMCTLD LOG FOR DETAILS *** slurmstepd: error: *** JOB 14650 ON h100-st-p548xlarge-2 CANCELLED AT 2024-08-23T04:48:27 DUE TO NODE FAILURE, SEE SLURMCTLD LOG FOR DETAILS *** Loading checkpoint shards: 0%| | 0/4 [00:00 main(eval_args, slurm_args) File "/opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/mmassist/eval/eval.py", line 154, in main job.results() # wait for the job to finish File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in results return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 287, in return [tp.cast(R, sub_job.result()) for sub_job in self._sub_jobs] File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 266, in result r = self.results() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 289, in results outcome, result = self._get_outcome_and_result() File "/data/home/imzyc/miniconda3/envs/mm/lib/python3.10/site-packages/submitit/core/core.py", line 384, in _get_outcome_and_result raise utils.UncompletedJobError("\n".join(message)) submitit.core.utils.UncompletedJobError: Job 14932 (task: 0) with path /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14932/14932_0_result.pkl has not produced any output (state: NODE_FAIL) No output/error stream produced ! Check: /opt/hpcaas/.mounts/fs-036153e63d56f4dc2/home/imzyc/project/proactive-assist/slurm_logs/14932/14932_0_log.out Model: /fsx_0/user/imzyc/proact_exps/20240822-L4096-I5-ep4-NOSEP-nr0.1-klgmix-1s-lora-bs384-debug {'assembly101/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}]}, 'ego4d/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}]}, 'egoexolearn/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.4}]}, 'epickitchens/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.2}]}, 'holoassist/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.3}]}, 'wtag/dialog-klg_test_L0_I5': {'stream': [{'context_handling_method': 'summarize_and_drop', 'eval_max_seq_len': 4096, 'eval_max_seq_len_str': '4k', 'inference_runner_type': 'stream', 'not_talk_threshold': 0.5}]}} Evaluation datasets: * ego4d/dialog-klg_test | num samples: 99 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.3 Evalulation: ego4d-dialog-klg_test_L0_I5/stream/notalk0.3-maxlen_4k Metrics: jaccard_index: 0.2215 missing_rate: 0.4983 redundant_rate: 0.5262 semantic_score: 0.6923 time_diff: 1.5123 precision: 0.3258 recall: 0.3450 F1: 0.3351 num_matched: 1741.0000 num_mismatched: 791.0000 num_missed: 2515.0000 num_redundant: 2812.0000 Bleu_1: 0.3835 Bleu_1_w: 0.0850 Bleu_2: 0.2653 Bleu_2_w: 0.0588 Bleu_3: 0.1941 Bleu_3_w: 0.0430 Bleu_4: 0.1485 Bleu_4_w: 0.0329 CIDEr: 0.8355 CIDEr_w: 0.1851 METEOR: 0.1879 METEOR_w: 0.0416 Evaluation datasets: * epickitchens/dialog-klg_test | num samples: 150 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.2 Evalulation: epickitchens-dialog-klg_test_L0_I5/stream/notalk0.2-maxlen_4k Metrics: jaccard_index: 0.2144 missing_rate: 0.5547 redundant_rate: 0.3526 semantic_score: 0.6676 time_diff: 0.5445 precision: 0.3873 recall: 0.2664 F1: 0.3157 num_matched: 1607.0000 num_mismatched: 1079.0000 num_missed: 3346.0000 num_redundant: 1463.0000 Bleu_1: 0.3977 Bleu_1_w: 0.0853 Bleu_2: 0.2734 Bleu_2_w: 0.0586 Bleu_3: 0.2000 Bleu_3_w: 0.0429 Bleu_4: 0.1550 Bleu_4_w: 0.0332 CIDEr: 1.0101 CIDEr_w: 0.2166 METEOR: 0.1894 METEOR_w: 0.0406 Evaluation datasets: * holoassist/dialog-klg_test | num samples: 291 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.3 Evalulation: holoassist-dialog-klg_test_L0_I5/stream/notalk0.3-maxlen_4k Metrics: jaccard_index: 0.2842 missing_rate: 0.5910 redundant_rate: 0.0834 semantic_score: 0.7066 time_diff: 0.2819 precision: 0.6606 recall: 0.2948 F1: 0.4076 num_matched: 4105.0000 num_mismatched: 1591.0000 num_missed: 8230.0000 num_redundant: 518.0000 Bleu_1: 0.4468 Bleu_1_w: 0.1270 Bleu_2: 0.3305 Bleu_2_w: 0.0939 Bleu_3: 0.2574 Bleu_3_w: 0.0731 Bleu_4: 0.2054 Bleu_4_w: 0.0584 CIDEr: 1.3007 CIDEr_w: 0.3696 METEOR: 0.2151 METEOR_w: 0.0611 Evaluation datasets: * egoexolearn/dialog-klg_test | num samples: 123 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.4 Evalulation: egoexolearn-dialog-klg_test_L0_I5/stream/notalk0.4-maxlen_4k Metrics: jaccard_index: 0.1634 missing_rate: 0.6318 redundant_rate: 0.5117 semantic_score: 0.6693 time_diff: 0.6571 precision: 0.3003 recall: 0.2264 F1: 0.2582 num_matched: 2730.0000 num_mismatched: 1710.0000 num_missed: 7618.0000 num_redundant: 4652.0000 Bleu_1: 0.3995 Bleu_1_w: 0.0653 Bleu_2: 0.2734 Bleu_2_w: 0.0447 Bleu_3: 0.1992 Bleu_3_w: 0.0325 Bleu_4: 0.1510 Bleu_4_w: 0.0247 CIDEr: 0.8387 CIDEr_w: 0.1370 METEOR: 0.1802 METEOR_w: 0.0294 Evaluation datasets: * assembly101/dialog-klg_test | num samples: 336 Updating eval setup: inference_runner_type: None -> stream Updating eval setup: not_talk_threshold: 0.5 -> 0.3 Evalulation: assembly101-dialog-klg_test_L0_I5/stream/notalk0.3-maxlen_4k Metrics: jaccard_index: 0.2835 missing_rate: 0.4738 redundant_rate: 0.2770 semantic_score: 0.7053 time_diff: 0.6322 precision: 0.4681 recall: 0.3407 F1: 0.3944 num_matched: 2814.0000 num_mismatched: 1532.0000 num_missed: 3914.0000 num_redundant: 1665.0000 Bleu_1: 0.4403 Bleu_1_w: 0.1249 Bleu_2: 0.3314 Bleu_2_w: 0.0939 Bleu_3: 0.2589 Bleu_3_w: 0.0734 Bleu_4: 0.2095 Bleu_4_w: 0.0594 CIDEr: 1.1329 CIDEr_w: 0.3212 METEOR: 0.2114 METEOR_w: 0.0599 Evaluation datasets: * wtag/dialog-klg_test | num samples: 21 Updating eval setup: inference_runner_type: None -> stream Evalulation: wtag-dialog-klg_test_L0_I5/stream/notalk0.5-maxlen_4k Metrics: jaccard_index: 0.2215 missing_rate: 0.3536 redundant_rate: 0.3931 semantic_score: 0.6719 time_diff: 1.4128 precision: 0.2950 recall: 0.3142 F1: 0.3043 num_matched: 367.0000 num_mismatched: 388.0000 num_missed: 413.0000 num_redundant: 489.0000 Bleu_1: 0.3966 Bleu_1_w: 0.0879 Bleu_2: 0.2880 Bleu_2_w: 0.0638 Bleu_3: 0.2202 Bleu_3_w: 0.0488 Bleu_4: 0.1728 Bleu_4_w: 0.0383 CIDEr: 1.2909 CIDEr_w: 0.2859 METEOR: 0.2019 METEOR_w: 0.0447 All Finished! Time: 24.86 minutes Model: /fsx_0/user/imzyc/proact_exps/20240822-L4096-I5-ep4-NOSEP-nr0.1-klgmix-1s-lora-bs384-debug Runs: ego4d/dialog-klg_test_L0_I5|stream|4k|0.3|summarize_and_drop epickitchens/dialog-klg_test_L0_I5|stream|4k|0.2|summarize_and_drop holoassist/dialog-klg_test_L0_I5|stream|4k|0.3|summarize_and_drop egoexolearn/dialog-klg_test_L0_I5|stream|4k|0.4|summarize_and_drop assembly101/dialog-klg_test_L0_I5|stream|4k|0.3|summarize_and_drop wtag/dialog-klg_test_L0_I5|stream|4k|0.5|summarize_and_drop