rajveerb Claude Haiku 4.5 commited on
Commit
299e4c2
·
1 Parent(s): ee6a98b

Remove oeda_pbas crawler run logs from H200 and H100

Browse files

This scheduler is not included in the paper evaluation.

Co-Authored-By: Claude Haiku 4.5 <noreply@anthropic.com>

Files changed (24) hide show
  1. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/config_10-01-2025_01-15-09.yaml +0 -29
  2. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/run_metrics.csv +0 -3
  3. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/config_10-01-2025_03-09-15.yaml +0 -29
  4. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/run_metrics.csv +0 -3
  5. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/config_10-01-2025_05-17-42.yaml +0 -29
  6. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/run_metrics.csv +0 -3
  7. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/config_10-01-2025_08-38-53.yaml +0 -29
  8. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/run_metrics.csv +0 -3
  9. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/config_10-01-2025_15-12-23.yaml +0 -29
  10. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/run_metrics.csv +0 -3
  11. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/config_10-02-2025_03-51-09.yaml +0 -29
  12. run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/run_metrics.csv +0 -3
  13. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/config_09-28-2025_07-08-36.yaml +0 -29
  14. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/run_metrics.csv +0 -3
  15. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/config_09-28-2025_08-54-10.yaml +0 -29
  16. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/run_metrics.csv +0 -3
  17. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/config_09-28-2025_11-00-03.yaml +0 -29
  18. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/run_metrics.csv +0 -3
  19. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/config_09-28-2025_14-20-54.yaml +0 -29
  20. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/run_metrics.csv +0 -3
  21. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/config_09-28-2025_20-48-11.yaml +0 -29
  22. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/run_metrics.csv +0 -3
  23. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/config_09-29-2025_09-22-44.yaml +0 -29
  24. run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/run_metrics.csv +0 -3
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/config_10-01-2025_01-15-09.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.0625
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_0625.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b31ae2ebdfb17d1f1f0ac78145b20922141b36baeb6798db720fde9f705b97b9
3
- size 8386977
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/config_10-01-2025_03-09-15.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.125
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_125.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9a4cf9978d7f74f436a1c40580384a9ab7ef610e573da5dc4d3b0b68be444d3
3
- size 11060331
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/config_10-01-2025_05-17-42.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.25
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_25.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c39f3db67d54c3d5465e0e959fe3271ea3f2a36bbb738e36c6792fac8bf1cd46
3
- size 12741366
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/config_10-01-2025_08-38-53.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.5
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_5.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:107435c62aab8e6a06d742eac2217811b03e3ce2c33edd6b0de1fcfdb2b5ad2c
3
- size 12760787
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/config_10-01-2025_15-12-23.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 1
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_1.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b90f096b6b341813c1c1fe8fd22881c27c8be30c96e260053b3484ca58fb1951
3
- size 12752956
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/config_10-02-2025_03-51-09.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 2
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_2.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:defcc978de4adee7d883e6b47c39018ce9d655ab66b672c8ec23a61b6e25ba3a
3
- size 12713960
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/config_09-28-2025_07-08-36.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.0625
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_0625.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2edbd26c6d1c83fc9d81f15db5fb35a56bc766ee6ff83d4d3cd119bf0ff072f0
3
- size 9225582
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/config_09-28-2025_08-54-10.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.125
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_125.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc3bf05f99e51d8974bb87057ad67db0858865c2af3967aefe56fc69265ce50c
3
- size 12083902
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/config_09-28-2025_11-00-03.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.25
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_25.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:67275890612d7638b7f17c9feb89db971e9454075697877a38a630ea9a1df32c
3
- size 12780640
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/config_09-28-2025_14-20-54.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 0.5
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_5.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c39cb63514d2335dfba974a715c5daf70e9b48f5c001e6ab19787bbc5c63377b
3
- size 12819805
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/config_09-28-2025_20-48-11.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 1
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_1.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e308f9cf0c9714032635ff174eb1be903d2f77c6f3614379009fcd9d45d27edc
3
- size 12765843
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/config_09-29-2025_09-22-44.yaml DELETED
@@ -1,29 +0,0 @@
1
- config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
2
- experiments:
3
- - concurrency: true
4
- stream: true
5
- log_stats: true
6
- model:
7
- gpu_memory_utilization: 0.8
8
- max_token_budget: 2048
9
- name: meta-llama/Meta-Llama-3.1-8B-Instruct
10
- tensor_parallel_size: 2
11
- prefix_caching_flag: false
12
- replay:
13
- end_template: "\nQuestion: \n"
14
- max_output_tokens: 1
15
- poisson_avg_arrival_time: 2
16
- template: 'Answer the question based only on the provided context. If the answer
17
- is not in the context, say "I don''t know."
18
-
19
- Context:
20
-
21
- '
22
- scheduler: oeda_pbas
23
- trace_dir: data/crawl/traces/simpleQA_ALL
24
- use_recomputation_latency_predictor: 1
25
- use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_2.txt
26
- use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
27
- use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
28
- use_swap_kernel: 1
29
- use_swap_latency_predictor: 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/run_metrics.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5c440f37ff81d97c7147a25760acea230c265f1887b0465ef2664acba019971
3
- size 12706352