rajveerb Claude Haiku 4.5 commited on
Commit ·
299e4c2
1
Parent(s): ee6a98b
Remove oeda_pbas crawler run logs from H200 and H100
Browse filesThis scheduler is not included in the paper evaluation.
Co-Authored-By: Claude Haiku 4.5 <noreply@anthropic.com>
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/config_10-01-2025_01-15-09.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/run_metrics.csv +0 -3
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/config_10-01-2025_03-09-15.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/run_metrics.csv +0 -3
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/config_10-01-2025_05-17-42.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/run_metrics.csv +0 -3
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/config_10-01-2025_08-38-53.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/run_metrics.csv +0 -3
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/config_10-01-2025_15-12-23.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/run_metrics.csv +0 -3
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/config_10-02-2025_03-51-09.yaml +0 -29
- run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/config_09-28-2025_07-08-36.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/config_09-28-2025_08-54-10.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/config_09-28-2025_11-00-03.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/config_09-28-2025_14-20-54.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/config_09-28-2025_20-48-11.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/run_metrics.csv +0 -3
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/config_09-29-2025_09-22-44.yaml +0 -29
- run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/run_metrics.csv +0 -3
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/config_10-01-2025_01-15-09.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.0625
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_0625.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_01-15-09/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:b31ae2ebdfb17d1f1f0ac78145b20922141b36baeb6798db720fde9f705b97b9
|
| 3 |
-
size 8386977
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/config_10-01-2025_03-09-15.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.125
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_125.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_03-09-15/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:e9a4cf9978d7f74f436a1c40580384a9ab7ef610e573da5dc4d3b0b68be444d3
|
| 3 |
-
size 11060331
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/config_10-01-2025_05-17-42.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.25
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_25.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_05-17-42/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c39f3db67d54c3d5465e0e959fe3271ea3f2a36bbb738e36c6792fac8bf1cd46
|
| 3 |
-
size 12741366
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/config_10-01-2025_08-38-53.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.5
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_0_5.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_08-38-53/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:107435c62aab8e6a06d742eac2217811b03e3ce2c33edd6b0de1fcfdb2b5ad2c
|
| 3 |
-
size 12760787
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/config_10-01-2025_15-12-23.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 1
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_1.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-01-2025_15-12-23/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:b90f096b6b341813c1c1fe8fd22881c27c8be30c96e260053b3484ca58fb1951
|
| 3 |
-
size 12752956
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/config_10-02-2025_03-51-09.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H100_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 2
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H100_enhanced_schedulers_v1_full/poisson_delays_2.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H100_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H100_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H100_enhanced_schedulers_v1_full/oeda_pbas/10-02-2025_03-51-09/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:defcc978de4adee7d883e6b47c39018ce9d655ab66b672c8ec23a61b6e25ba3a
|
| 3 |
-
size 12713960
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/config_09-28-2025_07-08-36.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.0625
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_0625.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_07-08-36/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:2edbd26c6d1c83fc9d81f15db5fb35a56bc766ee6ff83d4d3cd119bf0ff072f0
|
| 3 |
-
size 9225582
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/config_09-28-2025_08-54-10.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.125
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_125.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_08-54-10/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:fc3bf05f99e51d8974bb87057ad67db0858865c2af3967aefe56fc69265ce50c
|
| 3 |
-
size 12083902
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/config_09-28-2025_11-00-03.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.25
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_25.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_11-00-03/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:67275890612d7638b7f17c9feb89db971e9454075697877a38a630ea9a1df32c
|
| 3 |
-
size 12780640
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/config_09-28-2025_14-20-54.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 0.5
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_0_5.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_14-20-54/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c39cb63514d2335dfba974a715c5daf70e9b48f5c001e6ab19787bbc5c63377b
|
| 3 |
-
size 12819805
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/config_09-28-2025_20-48-11.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 1
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_1.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-28-2025_20-48-11/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:e308f9cf0c9714032635ff174eb1be903d2f77c6f3614379009fcd9d45d27edc
|
| 3 |
-
size 12765843
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/config_09-29-2025_09-22-44.yaml
DELETED
|
@@ -1,29 +0,0 @@
|
|
| 1 |
-
config_for_run: driver/crawler/run_log/H200_enhanced_schedulers_v1_full/oeda_pbas
|
| 2 |
-
experiments:
|
| 3 |
-
- concurrency: true
|
| 4 |
-
stream: true
|
| 5 |
-
log_stats: true
|
| 6 |
-
model:
|
| 7 |
-
gpu_memory_utilization: 0.8
|
| 8 |
-
max_token_budget: 2048
|
| 9 |
-
name: meta-llama/Meta-Llama-3.1-8B-Instruct
|
| 10 |
-
tensor_parallel_size: 2
|
| 11 |
-
prefix_caching_flag: false
|
| 12 |
-
replay:
|
| 13 |
-
end_template: "\nQuestion: \n"
|
| 14 |
-
max_output_tokens: 1
|
| 15 |
-
poisson_avg_arrival_time: 2
|
| 16 |
-
template: 'Answer the question based only on the provided context. If the answer
|
| 17 |
-
is not in the context, say "I don''t know."
|
| 18 |
-
|
| 19 |
-
Context:
|
| 20 |
-
|
| 21 |
-
'
|
| 22 |
-
scheduler: oeda_pbas
|
| 23 |
-
trace_dir: data/crawl/traces/simpleQA_ALL
|
| 24 |
-
use_recomputation_latency_predictor: 1
|
| 25 |
-
use_saved_poisson_delays: driver/crawler/configs/H200_enhanced_schedulers_v1_full/poisson_delays_2.txt
|
| 26 |
-
use_saved_recomputation_data_path: driver/recomputation_perf_model/H200_tp2_recomputation_latency.json
|
| 27 |
-
use_saved_swap_latency_data_path: driver/swap_perf_model/H200_tp2_swap_kernel_latency.json
|
| 28 |
-
use_swap_kernel: 1
|
| 29 |
-
use_swap_latency_predictor: 1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_log/crawler/H200_enhanced_schedulers_v1_full/oeda_pbas/09-29-2025_09-22-44/run_metrics.csv
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c5c440f37ff81d97c7147a25760acea230c265f1887b0465ef2664acba019971
|
| 3 |
-
size 12706352
|
|
|
|
|
|
|
|
|
|
|
|