Upload batch 17/20 (100 files)
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/config.json +50 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/phase_transitions.json +644 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/plots/gpu_metrics.png +3 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/plots/gpu_power_interactive.html +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/gpu_samples.csv +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/inference_results.csv +68 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/phase_power_summary.csv +10 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/config.json +50 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json +578 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv +96 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv +7 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/config.json +50 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json +578 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png +3 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv +71 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv +7 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/config.json +50 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json +578 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png +3 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv +100 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv +7 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/config.json +50 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json +578 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png +3 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv +0 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv +91 -0
- outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv +8 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/phase_transitions.json +68 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/transition_points.csv +1 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/config.json +52 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/phase_transitions.json +542 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/transition_points.csv +1 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 3,
|
| 7 |
+
"train_max_steps": null,
|
| 8 |
+
"batch_size": 4,
|
| 9 |
+
"gradient_accumulation_steps": 4,
|
| 10 |
+
"learning_rate": 0.0002,
|
| 11 |
+
"lora_r": 16,
|
| 12 |
+
"lora_alpha": 32,
|
| 13 |
+
"lora_dropout": 0.05,
|
| 14 |
+
"max_seq_length": 512,
|
| 15 |
+
"use_4bit": false,
|
| 16 |
+
"use_bf16": true,
|
| 17 |
+
"inference_total_requests": 30,
|
| 18 |
+
"inference_interval_sec": 5.0,
|
| 19 |
+
"inference_pattern": "fixed",
|
| 20 |
+
"inference_variable_min": 2.0,
|
| 21 |
+
"inference_variable_max": 10.0,
|
| 22 |
+
"inference_burst_count": 3,
|
| 23 |
+
"inference_burst_gap": 1.0,
|
| 24 |
+
"inference_burst_pause": 15.0,
|
| 25 |
+
"gen_max_new_tokens": 128,
|
| 26 |
+
"gen_temperature": 0.7,
|
| 27 |
+
"gen_do_sample": true,
|
| 28 |
+
"prompt_set": "alpaca",
|
| 29 |
+
"clock_lock_gpu": null,
|
| 30 |
+
"clock_lock_mem": null,
|
| 31 |
+
"ramp_enabled": true,
|
| 32 |
+
"ramp_start_w": 80,
|
| 33 |
+
"ramp_end_w": 575,
|
| 34 |
+
"ramp_steps": 10,
|
| 35 |
+
"ramp_step_duration": 3.0,
|
| 36 |
+
"device": "cuda:0",
|
| 37 |
+
"gpu_index": 0,
|
| 38 |
+
"skip_training": false,
|
| 39 |
+
"skip_inference": false,
|
| 40 |
+
"model": "qwen3-4b",
|
| 41 |
+
"power_cap": null,
|
| 42 |
+
"gpu_info": {
|
| 43 |
+
"gpu_name": "NVIDIA RTX PRO 6000 Blackwell Workstation Edition",
|
| 44 |
+
"tdp_w": 600,
|
| 45 |
+
"memory_total_mb": 97887,
|
| 46 |
+
"driver_version": "580.82.07",
|
| 47 |
+
"cuda_version": "13.0"
|
| 48 |
+
},
|
| 49 |
+
"experiment_timestamp": "2026-02-12T23:22:55.043180"
|
| 50 |
+
}
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/phase_transitions.json
ADDED
|
@@ -0,0 +1,644 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-12T23:10:27.791660",
|
| 4 |
+
"timestamp_ms": 1770905427791.6714,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-12T23:10:57.791893",
|
| 10 |
+
"timestamp_ms": 1770905457791.9326,
|
| 11 |
+
"phase": "ramp",
|
| 12 |
+
"sub_phase": "up_80W_to_575W"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-12T23:11:28.424992",
|
| 16 |
+
"timestamp_ms": 1770905488425.0273,
|
| 17 |
+
"phase": "ramp",
|
| 18 |
+
"sub_phase": "up_complete"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-12T23:11:28.425115",
|
| 22 |
+
"timestamp_ms": 1770905488425.1223,
|
| 23 |
+
"phase": "train",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-12T23:11:34.950501",
|
| 28 |
+
"timestamp_ms": 1770905494950.513,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "epoch_1_start"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-12T23:13:53.933593",
|
| 34 |
+
"timestamp_ms": 1770905633933.6033,
|
| 35 |
+
"phase": "validation",
|
| 36 |
+
"sub_phase": "epoch_1"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-12T23:13:55.461875",
|
| 40 |
+
"timestamp_ms": 1770905635461.887,
|
| 41 |
+
"phase": "train",
|
| 42 |
+
"sub_phase": "epoch_2_start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-12T23:16:14.034965",
|
| 46 |
+
"timestamp_ms": 1770905774034.9766,
|
| 47 |
+
"phase": "validation",
|
| 48 |
+
"sub_phase": "epoch_2"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-12T23:16:15.431776",
|
| 52 |
+
"timestamp_ms": 1770905775431.788,
|
| 53 |
+
"phase": "train",
|
| 54 |
+
"sub_phase": "epoch_3_start"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-12T23:18:34.220385",
|
| 58 |
+
"timestamp_ms": 1770905914220.3955,
|
| 59 |
+
"phase": "validation",
|
| 60 |
+
"sub_phase": "epoch_3"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-12T23:18:35.609607",
|
| 64 |
+
"timestamp_ms": 1770905915609.6187,
|
| 65 |
+
"phase": "train",
|
| 66 |
+
"sub_phase": "finished"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-12T23:18:35.617531",
|
| 70 |
+
"timestamp_ms": 1770905915617.538,
|
| 71 |
+
"phase": "idle_mid",
|
| 72 |
+
"sub_phase": ""
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-12T23:19:05.617730",
|
| 76 |
+
"timestamp_ms": 1770905945617.7683,
|
| 77 |
+
"phase": "inference",
|
| 78 |
+
"sub_phase": "start"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-12T23:19:09.242682",
|
| 82 |
+
"timestamp_ms": 1770905949242.6956,
|
| 83 |
+
"phase": "inference",
|
| 84 |
+
"sub_phase": "session_start_pattern_fixed"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-12T23:19:09.242830",
|
| 88 |
+
"timestamp_ms": 1770905949242.8376,
|
| 89 |
+
"phase": "inference",
|
| 90 |
+
"sub_phase": "req_0_prefill"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-12T23:19:09.292048",
|
| 94 |
+
"timestamp_ms": 1770905949292.0583,
|
| 95 |
+
"phase": "inference",
|
| 96 |
+
"sub_phase": "req_0_decode"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-12T23:19:09.413558",
|
| 100 |
+
"timestamp_ms": 1770905949413.5671,
|
| 101 |
+
"phase": "inference",
|
| 102 |
+
"sub_phase": "req_0_idle_between"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-12T23:19:14.413900",
|
| 106 |
+
"timestamp_ms": 1770905954413.9333,
|
| 107 |
+
"phase": "inference",
|
| 108 |
+
"sub_phase": "req_1_prefill"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-12T23:19:14.446453",
|
| 112 |
+
"timestamp_ms": 1770905954446.461,
|
| 113 |
+
"phase": "inference",
|
| 114 |
+
"sub_phase": "req_1_decode"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-12T23:19:14.475281",
|
| 118 |
+
"timestamp_ms": 1770905954475.2888,
|
| 119 |
+
"phase": "inference",
|
| 120 |
+
"sub_phase": "req_1_idle_between"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-12T23:19:19.475613",
|
| 124 |
+
"timestamp_ms": 1770905959475.6492,
|
| 125 |
+
"phase": "inference",
|
| 126 |
+
"sub_phase": "req_2_prefill"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-12T23:19:19.511793",
|
| 130 |
+
"timestamp_ms": 1770905959511.8044,
|
| 131 |
+
"phase": "inference",
|
| 132 |
+
"sub_phase": "req_2_decode"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-12T23:19:19.543025",
|
| 136 |
+
"timestamp_ms": 1770905959543.0342,
|
| 137 |
+
"phase": "inference",
|
| 138 |
+
"sub_phase": "req_2_idle_between"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-12T23:19:24.543305",
|
| 142 |
+
"timestamp_ms": 1770905964543.3376,
|
| 143 |
+
"phase": "inference",
|
| 144 |
+
"sub_phase": "req_3_prefill"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-12T23:19:24.577655",
|
| 148 |
+
"timestamp_ms": 1770905964577.6653,
|
| 149 |
+
"phase": "inference",
|
| 150 |
+
"sub_phase": "req_3_decode"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-12T23:19:25.670728",
|
| 154 |
+
"timestamp_ms": 1770905965670.7363,
|
| 155 |
+
"phase": "inference",
|
| 156 |
+
"sub_phase": "req_3_idle_between"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-12T23:19:30.671047",
|
| 160 |
+
"timestamp_ms": 1770905970671.0818,
|
| 161 |
+
"phase": "inference",
|
| 162 |
+
"sub_phase": "req_4_prefill"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-12T23:19:30.705384",
|
| 166 |
+
"timestamp_ms": 1770905970705.3943,
|
| 167 |
+
"phase": "inference",
|
| 168 |
+
"sub_phase": "req_4_decode"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-12T23:19:33.163296",
|
| 172 |
+
"timestamp_ms": 1770905973163.3047,
|
| 173 |
+
"phase": "inference",
|
| 174 |
+
"sub_phase": "req_4_idle_between"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-12T23:19:38.163572",
|
| 178 |
+
"timestamp_ms": 1770905978163.6067,
|
| 179 |
+
"phase": "inference",
|
| 180 |
+
"sub_phase": "req_5_prefill"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-12T23:19:38.198197",
|
| 184 |
+
"timestamp_ms": 1770905978198.2083,
|
| 185 |
+
"phase": "inference",
|
| 186 |
+
"sub_phase": "req_5_decode"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-12T23:19:38.228310",
|
| 190 |
+
"timestamp_ms": 1770905978228.318,
|
| 191 |
+
"phase": "inference",
|
| 192 |
+
"sub_phase": "req_5_idle_between"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-12T23:19:43.228630",
|
| 196 |
+
"timestamp_ms": 1770905983228.666,
|
| 197 |
+
"phase": "inference",
|
| 198 |
+
"sub_phase": "req_6_prefill"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-12T23:19:43.264009",
|
| 202 |
+
"timestamp_ms": 1770905983264.02,
|
| 203 |
+
"phase": "inference",
|
| 204 |
+
"sub_phase": "req_6_decode"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-12T23:19:43.294105",
|
| 208 |
+
"timestamp_ms": 1770905983294.1143,
|
| 209 |
+
"phase": "inference",
|
| 210 |
+
"sub_phase": "req_6_idle_between"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-12T23:19:48.294396",
|
| 214 |
+
"timestamp_ms": 1770905988294.4314,
|
| 215 |
+
"phase": "inference",
|
| 216 |
+
"sub_phase": "req_7_prefill"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-12T23:19:48.330130",
|
| 220 |
+
"timestamp_ms": 1770905988330.1404,
|
| 221 |
+
"phase": "inference",
|
| 222 |
+
"sub_phase": "req_7_decode"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-12T23:19:48.696385",
|
| 226 |
+
"timestamp_ms": 1770905988696.3933,
|
| 227 |
+
"phase": "inference",
|
| 228 |
+
"sub_phase": "req_7_idle_between"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-12T23:19:53.696665",
|
| 232 |
+
"timestamp_ms": 1770905993696.7004,
|
| 233 |
+
"phase": "inference",
|
| 234 |
+
"sub_phase": "req_8_prefill"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-12T23:19:53.730256",
|
| 238 |
+
"timestamp_ms": 1770905993730.2666,
|
| 239 |
+
"phase": "inference",
|
| 240 |
+
"sub_phase": "req_8_decode"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-12T23:19:55.126181",
|
| 244 |
+
"timestamp_ms": 1770905995126.189,
|
| 245 |
+
"phase": "inference",
|
| 246 |
+
"sub_phase": "req_8_idle_between"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-12T23:20:00.126490",
|
| 250 |
+
"timestamp_ms": 1770906000126.525,
|
| 251 |
+
"phase": "inference",
|
| 252 |
+
"sub_phase": "req_9_prefill"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-12T23:20:00.161691",
|
| 256 |
+
"timestamp_ms": 1770906000161.701,
|
| 257 |
+
"phase": "inference",
|
| 258 |
+
"sub_phase": "req_9_decode"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-12T23:20:00.193400",
|
| 262 |
+
"timestamp_ms": 1770906000193.4082,
|
| 263 |
+
"phase": "inference",
|
| 264 |
+
"sub_phase": "req_9_idle_between"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-12T23:20:05.193725",
|
| 268 |
+
"timestamp_ms": 1770906005193.7622,
|
| 269 |
+
"phase": "inference",
|
| 270 |
+
"sub_phase": "req_10_prefill"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-12T23:20:05.227539",
|
| 274 |
+
"timestamp_ms": 1770906005227.5496,
|
| 275 |
+
"phase": "inference",
|
| 276 |
+
"sub_phase": "req_10_decode"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-12T23:20:05.260873",
|
| 280 |
+
"timestamp_ms": 1770906005260.8816,
|
| 281 |
+
"phase": "inference",
|
| 282 |
+
"sub_phase": "req_10_idle_between"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-12T23:20:10.261201",
|
| 286 |
+
"timestamp_ms": 1770906010261.2363,
|
| 287 |
+
"phase": "inference",
|
| 288 |
+
"sub_phase": "req_11_prefill"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-12T23:20:10.295734",
|
| 292 |
+
"timestamp_ms": 1770906010295.7441,
|
| 293 |
+
"phase": "inference",
|
| 294 |
+
"sub_phase": "req_11_decode"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-12T23:20:10.333269",
|
| 298 |
+
"timestamp_ms": 1770906010333.2773,
|
| 299 |
+
"phase": "inference",
|
| 300 |
+
"sub_phase": "req_11_idle_between"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-12T23:20:15.333552",
|
| 304 |
+
"timestamp_ms": 1770906015333.5867,
|
| 305 |
+
"phase": "inference",
|
| 306 |
+
"sub_phase": "req_12_prefill"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-12T23:20:15.368683",
|
| 310 |
+
"timestamp_ms": 1770906015368.6934,
|
| 311 |
+
"phase": "inference",
|
| 312 |
+
"sub_phase": "req_12_decode"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-12T23:20:15.825799",
|
| 316 |
+
"timestamp_ms": 1770906015825.8066,
|
| 317 |
+
"phase": "inference",
|
| 318 |
+
"sub_phase": "req_12_idle_between"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-12T23:20:20.826097",
|
| 322 |
+
"timestamp_ms": 1770906020826.1335,
|
| 323 |
+
"phase": "inference",
|
| 324 |
+
"sub_phase": "req_13_prefill"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-12T23:20:20.863892",
|
| 328 |
+
"timestamp_ms": 1770906020863.9033,
|
| 329 |
+
"phase": "inference",
|
| 330 |
+
"sub_phase": "req_13_decode"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-12T23:20:21.208101",
|
| 334 |
+
"timestamp_ms": 1770906021208.1091,
|
| 335 |
+
"phase": "inference",
|
| 336 |
+
"sub_phase": "req_13_idle_between"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-12T23:20:26.208403",
|
| 340 |
+
"timestamp_ms": 1770906026208.4373,
|
| 341 |
+
"phase": "inference",
|
| 342 |
+
"sub_phase": "req_14_prefill"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-12T23:20:26.246710",
|
| 346 |
+
"timestamp_ms": 1770906026246.72,
|
| 347 |
+
"phase": "inference",
|
| 348 |
+
"sub_phase": "req_14_decode"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-12T23:20:26.276471",
|
| 352 |
+
"timestamp_ms": 1770906026276.4795,
|
| 353 |
+
"phase": "inference",
|
| 354 |
+
"sub_phase": "req_14_idle_between"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-12T23:20:31.276732",
|
| 358 |
+
"timestamp_ms": 1770906031276.765,
|
| 359 |
+
"phase": "inference",
|
| 360 |
+
"sub_phase": "req_15_prefill"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-12T23:20:31.313998",
|
| 364 |
+
"timestamp_ms": 1770906031314.0073,
|
| 365 |
+
"phase": "inference",
|
| 366 |
+
"sub_phase": "req_15_decode"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-12T23:20:31.343391",
|
| 370 |
+
"timestamp_ms": 1770906031343.3984,
|
| 371 |
+
"phase": "inference",
|
| 372 |
+
"sub_phase": "req_15_idle_between"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-12T23:20:36.343695",
|
| 376 |
+
"timestamp_ms": 1770906036343.7302,
|
| 377 |
+
"phase": "inference",
|
| 378 |
+
"sub_phase": "req_16_prefill"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-12T23:20:36.382149",
|
| 382 |
+
"timestamp_ms": 1770906036382.1584,
|
| 383 |
+
"phase": "inference",
|
| 384 |
+
"sub_phase": "req_16_decode"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-12T23:20:38.980482",
|
| 388 |
+
"timestamp_ms": 1770906038980.4905,
|
| 389 |
+
"phase": "inference",
|
| 390 |
+
"sub_phase": "req_16_idle_between"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-12T23:20:43.980793",
|
| 394 |
+
"timestamp_ms": 1770906043980.827,
|
| 395 |
+
"phase": "inference",
|
| 396 |
+
"sub_phase": "req_17_prefill"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-12T23:20:44.015368",
|
| 400 |
+
"timestamp_ms": 1770906044015.3777,
|
| 401 |
+
"phase": "inference",
|
| 402 |
+
"sub_phase": "req_17_decode"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-12T23:20:44.047052",
|
| 406 |
+
"timestamp_ms": 1770906044047.0608,
|
| 407 |
+
"phase": "inference",
|
| 408 |
+
"sub_phase": "req_17_idle_between"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-12T23:20:49.047389",
|
| 412 |
+
"timestamp_ms": 1770906049047.4285,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "req_18_prefill"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-12T23:20:49.094511",
|
| 418 |
+
"timestamp_ms": 1770906049094.5244,
|
| 419 |
+
"phase": "inference",
|
| 420 |
+
"sub_phase": "req_18_decode"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-12T23:20:49.592172",
|
| 424 |
+
"timestamp_ms": 1770906049592.1797,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "req_18_idle_between"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-12T23:20:54.592440",
|
| 430 |
+
"timestamp_ms": 1770906054592.4736,
|
| 431 |
+
"phase": "inference",
|
| 432 |
+
"sub_phase": "req_19_prefill"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-12T23:20:54.625974",
|
| 436 |
+
"timestamp_ms": 1770906054625.984,
|
| 437 |
+
"phase": "inference",
|
| 438 |
+
"sub_phase": "req_19_decode"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-12T23:20:57.257887",
|
| 442 |
+
"timestamp_ms": 1770906057257.897,
|
| 443 |
+
"phase": "inference",
|
| 444 |
+
"sub_phase": "req_19_idle_between"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-12T23:21:02.258180",
|
| 448 |
+
"timestamp_ms": 1770906062258.2158,
|
| 449 |
+
"phase": "inference",
|
| 450 |
+
"sub_phase": "req_20_prefill"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-12T23:21:02.296179",
|
| 454 |
+
"timestamp_ms": 1770906062296.1895,
|
| 455 |
+
"phase": "inference",
|
| 456 |
+
"sub_phase": "req_20_decode"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-12T23:21:02.328564",
|
| 460 |
+
"timestamp_ms": 1770906062328.5725,
|
| 461 |
+
"phase": "inference",
|
| 462 |
+
"sub_phase": "req_20_idle_between"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-12T23:21:07.328861",
|
| 466 |
+
"timestamp_ms": 1770906067328.9016,
|
| 467 |
+
"phase": "inference",
|
| 468 |
+
"sub_phase": "req_21_prefill"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-12T23:21:07.366006",
|
| 472 |
+
"timestamp_ms": 1770906067366.017,
|
| 473 |
+
"phase": "inference",
|
| 474 |
+
"sub_phase": "req_21_decode"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-12T23:21:07.397549",
|
| 478 |
+
"timestamp_ms": 1770906067397.558,
|
| 479 |
+
"phase": "inference",
|
| 480 |
+
"sub_phase": "req_21_idle_between"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-12T23:21:12.397822",
|
| 484 |
+
"timestamp_ms": 1770906072397.8887,
|
| 485 |
+
"phase": "inference",
|
| 486 |
+
"sub_phase": "req_22_prefill"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-12T23:21:12.433014",
|
| 490 |
+
"timestamp_ms": 1770906072433.0244,
|
| 491 |
+
"phase": "inference",
|
| 492 |
+
"sub_phase": "req_22_decode"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-12T23:21:12.481980",
|
| 496 |
+
"timestamp_ms": 1770906072482.0105,
|
| 497 |
+
"phase": "inference",
|
| 498 |
+
"sub_phase": "req_22_idle_between"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-12T23:21:17.482364",
|
| 502 |
+
"timestamp_ms": 1770906077482.3992,
|
| 503 |
+
"phase": "inference",
|
| 504 |
+
"sub_phase": "req_23_prefill"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-12T23:21:17.518886",
|
| 508 |
+
"timestamp_ms": 1770906077518.8962,
|
| 509 |
+
"phase": "inference",
|
| 510 |
+
"sub_phase": "req_23_decode"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-12T23:21:17.938490",
|
| 514 |
+
"timestamp_ms": 1770906077938.4973,
|
| 515 |
+
"phase": "inference",
|
| 516 |
+
"sub_phase": "req_23_idle_between"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-12T23:21:22.938780",
|
| 520 |
+
"timestamp_ms": 1770906082938.8147,
|
| 521 |
+
"phase": "inference",
|
| 522 |
+
"sub_phase": "req_24_prefill"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-12T23:21:22.973635",
|
| 526 |
+
"timestamp_ms": 1770906082973.6458,
|
| 527 |
+
"phase": "inference",
|
| 528 |
+
"sub_phase": "req_24_decode"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-12T23:21:24.826277",
|
| 532 |
+
"timestamp_ms": 1770906084826.2847,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "req_24_idle_between"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-12T23:21:29.826721",
|
| 538 |
+
"timestamp_ms": 1770906089826.7568,
|
| 539 |
+
"phase": "inference",
|
| 540 |
+
"sub_phase": "req_25_prefill"
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"timestamp": "2026-02-12T23:21:29.860331",
|
| 544 |
+
"timestamp_ms": 1770906089860.3416,
|
| 545 |
+
"phase": "inference",
|
| 546 |
+
"sub_phase": "req_25_decode"
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"timestamp": "2026-02-12T23:21:33.371541",
|
| 550 |
+
"timestamp_ms": 1770906093371.5503,
|
| 551 |
+
"phase": "inference",
|
| 552 |
+
"sub_phase": "req_25_idle_between"
|
| 553 |
+
},
|
| 554 |
+
{
|
| 555 |
+
"timestamp": "2026-02-12T23:21:38.371812",
|
| 556 |
+
"timestamp_ms": 1770906098371.8853,
|
| 557 |
+
"phase": "inference",
|
| 558 |
+
"sub_phase": "req_26_prefill"
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"timestamp": "2026-02-12T23:21:38.405564",
|
| 562 |
+
"timestamp_ms": 1770906098405.574,
|
| 563 |
+
"phase": "inference",
|
| 564 |
+
"sub_phase": "req_26_decode"
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"timestamp": "2026-02-12T23:21:38.435955",
|
| 568 |
+
"timestamp_ms": 1770906098435.9636,
|
| 569 |
+
"phase": "inference",
|
| 570 |
+
"sub_phase": "req_26_idle_between"
|
| 571 |
+
},
|
| 572 |
+
{
|
| 573 |
+
"timestamp": "2026-02-12T23:21:43.436277",
|
| 574 |
+
"timestamp_ms": 1770906103436.3127,
|
| 575 |
+
"phase": "inference",
|
| 576 |
+
"sub_phase": "req_27_prefill"
|
| 577 |
+
},
|
| 578 |
+
{
|
| 579 |
+
"timestamp": "2026-02-12T23:21:43.493709",
|
| 580 |
+
"timestamp_ms": 1770906103493.719,
|
| 581 |
+
"phase": "inference",
|
| 582 |
+
"sub_phase": "req_27_decode"
|
| 583 |
+
},
|
| 584 |
+
{
|
| 585 |
+
"timestamp": "2026-02-12T23:21:43.526771",
|
| 586 |
+
"timestamp_ms": 1770906103526.7803,
|
| 587 |
+
"phase": "inference",
|
| 588 |
+
"sub_phase": "req_27_idle_between"
|
| 589 |
+
},
|
| 590 |
+
{
|
| 591 |
+
"timestamp": "2026-02-12T23:21:48.527070",
|
| 592 |
+
"timestamp_ms": 1770906108527.107,
|
| 593 |
+
"phase": "inference",
|
| 594 |
+
"sub_phase": "req_28_prefill"
|
| 595 |
+
},
|
| 596 |
+
{
|
| 597 |
+
"timestamp": "2026-02-12T23:21:48.564559",
|
| 598 |
+
"timestamp_ms": 1770906108564.57,
|
| 599 |
+
"phase": "inference",
|
| 600 |
+
"sub_phase": "req_28_decode"
|
| 601 |
+
},
|
| 602 |
+
{
|
| 603 |
+
"timestamp": "2026-02-12T23:21:49.247344",
|
| 604 |
+
"timestamp_ms": 1770906109247.3518,
|
| 605 |
+
"phase": "inference",
|
| 606 |
+
"sub_phase": "req_28_idle_between"
|
| 607 |
+
},
|
| 608 |
+
{
|
| 609 |
+
"timestamp": "2026-02-12T23:21:54.247591",
|
| 610 |
+
"timestamp_ms": 1770906114247.6223,
|
| 611 |
+
"phase": "inference",
|
| 612 |
+
"sub_phase": "req_29_prefill"
|
| 613 |
+
},
|
| 614 |
+
{
|
| 615 |
+
"timestamp": "2026-02-12T23:21:54.278387",
|
| 616 |
+
"timestamp_ms": 1770906114278.3955,
|
| 617 |
+
"phase": "inference",
|
| 618 |
+
"sub_phase": "req_29_decode"
|
| 619 |
+
},
|
| 620 |
+
{
|
| 621 |
+
"timestamp": "2026-02-12T23:21:54.586277",
|
| 622 |
+
"timestamp_ms": 1770906114586.2847,
|
| 623 |
+
"phase": "inference",
|
| 624 |
+
"sub_phase": "session_finished"
|
| 625 |
+
},
|
| 626 |
+
{
|
| 627 |
+
"timestamp": "2026-02-12T23:21:54.586386",
|
| 628 |
+
"timestamp_ms": 1770906114586.3887,
|
| 629 |
+
"phase": "ramp",
|
| 630 |
+
"sub_phase": "down_575W_to_80W"
|
| 631 |
+
},
|
| 632 |
+
{
|
| 633 |
+
"timestamp": "2026-02-12T23:22:25.036127",
|
| 634 |
+
"timestamp_ms": 1770906145036.1602,
|
| 635 |
+
"phase": "ramp",
|
| 636 |
+
"sub_phase": "down_complete"
|
| 637 |
+
},
|
| 638 |
+
{
|
| 639 |
+
"timestamp": "2026-02-12T23:22:25.036240",
|
| 640 |
+
"timestamp_ms": 1770906145036.2478,
|
| 641 |
+
"phase": "idle_post",
|
| 642 |
+
"sub_phase": ""
|
| 643 |
+
}
|
| 644 |
+
]
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/inference_results.csv
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
request_idx,input_tokens,output_tokens,prefill_time_ms,decode_time_ms,prompt_preview,output_preview
|
| 2 |
+
0,21,2,36.01384162902832,121.16670608520508,"Evaluate this sentence for spelling and grammar mistakes
|
| 3 |
+
He finnished his meal a",.
|
| 4 |
+
1,7,1,31.063318252563477,28.5794734954834,Give two examples of a liquid.,
|
| 5 |
+
2,7,1,34.61647033691406,30.921220779418945,Outline the consequences of deforestation.,
|
| 6 |
+
3,14,39,32.773494720458984,1092.7789211273193,"Find the area of a circle given its radius.
|
| 7 |
+
Radius = 4","
|
| 8 |
+
Area = pi * (radius^2)
|
| 9 |
+
Area = pi * (4^2)
|
| 10 |
+
Area = pi * 16
|
| 11 |
+
Area = 50.27 (approx)"
|
| 12 |
+
4,11,90,32.85026550292969,2457.597017288208,Propose an ethical solution to the problem of data privacy," in the digital age.
|
| 13 |
+
|
| 14 |
+
One ethical solution to the problem of data privacy in the"
|
| 15 |
+
5,8,1,33.09035301208496,29.848575592041016,Design an app for a delivery company.,
|
| 16 |
+
6,12,1,33.872127532958984,29.80351448059082,Compare and contrast the Cuban Missile Crisis and the Vietnam War.,
|
| 17 |
+
7,7,13,34.12914276123047,365.9930229187012,What are the three primary colors?," The three primary colors are red, blue, and yellow."
|
| 18 |
+
8,14,51,31.924962997436523,1395.6196308135986,Generate a list of ten items a person might need for a camping trip,".
|
| 19 |
+
1. Tent
|
| 20 |
+
2. Sleeping bag
|
| 21 |
+
3. Cooking pot
|
| 22 |
+
4. Cooler
|
| 23 |
+
5. Headlamp
|
| 24 |
+
6. First aid kit
|
| 25 |
+
"
|
| 26 |
+
9,18,1,33.538103103637695,31.435728073120117,"Rewrite the following sentence using active voice.
|
| 27 |
+
The news report was read by t",
|
| 28 |
+
10,7,1,32.1955680847168,33.045291900634766,Describe the structure of an atom.,
|
| 29 |
+
11,7,1,32.987356185913086,31.778335571289062,Give three tips for staying healthy.,
|
| 30 |
+
12,12,16,33.51020812988281,456.8219184875488,Describe a time when you had to make a difficult decision.," What was the decision, and what factors did you consider in making it?"
|
| 31 |
+
13,46,12,35.60829162597656,343.9450263977051,"Extract the facts from the paragraph.
|
| 32 |
+
Online education continues to become more "," However, there are some disadvantages to online education, such"
|
| 33 |
+
14,7,1,36.79680824279785,29.503345489501953,What is the capital of France?,
|
| 34 |
+
15,9,1,35.75301170349121,29.148101806640625,Design a logo for a website about recycling.,
|
| 35 |
+
16,15,96,36.87763214111328,2598.0122089385986,"Write a short paragraph about the given topic.
|
| 36 |
+
The importance of using renewable"," sources.
|
| 37 |
+
|
| 38 |
+
Renewable energy sources are important for a number of reasons. Firs"
|
| 39 |
+
17,9,1,33.16211700439453,31.405210494995117,Generate a poem with 10 lines.,
|
| 40 |
+
18,21,18,45.11666297912598,497.387170791626,Arrange the words in the given sentence to form a grammatically correct sentence," over the lazy dog
|
| 41 |
+
Output: The quick brown fox jumped over the lazy dog."
|
| 42 |
+
19,11,97,32.019853591918945,2631.563901901245,Explain the use of word embeddings in Natural Language Processing,".
|
| 43 |
+
Word embeddings are a type of representation used in Natural Language Processi"
|
| 44 |
+
20,25,1,36.35525703430176,32.10711479187012,"Use the given data to calculate the median.
|
| 45 |
+
[2, 3, 7, 8, 10]",
|
| 46 |
+
21,11,1,35.39562225341797,31.244993209838867,Develop a plan to reduce electricity usage in a home.,
|
| 47 |
+
22,8,1,33.603668212890625,31.729698181152344,Provide one example for a cultural practice.,
|
| 48 |
+
23,13,15,33.695220947265625,419.2793369293213,"Transcribe the recording into text.
|
| 49 |
+
Recording about the history of Waterloo",":
|
| 50 |
+
The Battle of Waterloo, fought on June 18, "
|
| 51 |
+
24,6,68,33.350467681884766,1852.3433208465576,How did Julius Caesar die?, Julius Caesar was assassinated on the Ides of March (March 15th) in 44 BCE. He
|
| 52 |
+
25,13,128,32.04345703125,3510.875940322876,Identify the lines of longitude that divides North and South America.,"
|
| 53 |
+
|
| 54 |
+
A. 10°W and 20°W
|
| 55 |
+
B. 10°W and 15°W
|
| 56 |
+
C. 20°W and 40°W
|
| 57 |
+
D. 40°W and 60°W
|
| 58 |
+
Ans"
|
| 59 |
+
26,7,1,32.247066497802734,30.106782913208008,How can we reduce air pollution?,
|
| 60 |
+
27,66,1,55.09138107299805,32.77277946472168,"Rewrite the given paragraph in a shorter, easier to understand form.
|
| 61 |
+
Although it",
|
| 62 |
+
28,12,25,35.933732986450195,682.5323104858398,"Identify the odd one out.
|
| 63 |
+
Twitter, Instagram, Telegram","
|
| 64 |
+
Answer:
|
| 65 |
+
Telegram is the odd one out as it is a messaging app while Twitter and "
|
| 66 |
+
29,15,11,29.327869415283203,307.6364994049072,"Convert from celsius to fahrenheit.
|
| 67 |
+
Temperature in Celsius: 15","
|
| 68 |
+
Temperature in Fahrenheit: 59.0"
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/08_ramp_rate/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap_ramp/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3363,33.63002,1890.868126,56.225512,97.59,268.937,8.845
|
| 3 |
+
idle_post,3000,29.990052,273.011157,9.103412,14.599,14.747,7.648
|
| 4 |
+
idle_pre,3001,30.010068,1624.932625,54.146127,77.445,77.588,7.211
|
| 5 |
+
inference_decode,1922,19.219466,2716.792259,141.354562,210.075,210.751,8.845
|
| 6 |
+
inference_idle,14500,145.003192,13036.839324,89.907411,135.125,210.751,8.845
|
| 7 |
+
inference_prefill,112,1.11732,83.515622,74.752107,78.523,79.739,8.845
|
| 8 |
+
ramp,6715,67.599347,2342.344501,34.776654,79.077,98.107,7.013
|
| 9 |
+
train,41597,416.352758,234854.497605,564.07268,570.347,572.695,77.387
|
| 10 |
+
validation,432,4.315293,1793.497265,415.778065,570.144,570.144,196.661
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 3,
|
| 7 |
+
"train_max_steps": null,
|
| 8 |
+
"batch_size": 4,
|
| 9 |
+
"gradient_accumulation_steps": 4,
|
| 10 |
+
"learning_rate": 0.0002,
|
| 11 |
+
"lora_r": 16,
|
| 12 |
+
"lora_alpha": 32,
|
| 13 |
+
"lora_dropout": 0.05,
|
| 14 |
+
"max_seq_length": 512,
|
| 15 |
+
"use_4bit": false,
|
| 16 |
+
"use_bf16": true,
|
| 17 |
+
"inference_total_requests": 30,
|
| 18 |
+
"inference_interval_sec": 5.0,
|
| 19 |
+
"inference_pattern": "fixed",
|
| 20 |
+
"inference_variable_min": 2.0,
|
| 21 |
+
"inference_variable_max": 10.0,
|
| 22 |
+
"inference_burst_count": 3,
|
| 23 |
+
"inference_burst_gap": 1.0,
|
| 24 |
+
"inference_burst_pause": 15.0,
|
| 25 |
+
"gen_max_new_tokens": 128,
|
| 26 |
+
"gen_temperature": 0.7,
|
| 27 |
+
"gen_do_sample": true,
|
| 28 |
+
"prompt_set": "alpaca",
|
| 29 |
+
"clock_lock_gpu": null,
|
| 30 |
+
"clock_lock_mem": null,
|
| 31 |
+
"ramp_enabled": false,
|
| 32 |
+
"ramp_start_w": null,
|
| 33 |
+
"ramp_end_w": null,
|
| 34 |
+
"ramp_steps": 5,
|
| 35 |
+
"ramp_step_duration": 3.0,
|
| 36 |
+
"device": "cuda:0",
|
| 37 |
+
"gpu_index": 0,
|
| 38 |
+
"skip_training": true,
|
| 39 |
+
"skip_inference": false,
|
| 40 |
+
"model": "gpt2",
|
| 41 |
+
"power_cap": null,
|
| 42 |
+
"gpu_info": {
|
| 43 |
+
"gpu_name": "NVIDIA RTX PRO 6000 Blackwell Workstation Edition",
|
| 44 |
+
"tdp_w": 600,
|
| 45 |
+
"memory_total_mb": 97887,
|
| 46 |
+
"driver_version": "580.82.07",
|
| 47 |
+
"cuda_version": "13.0"
|
| 48 |
+
},
|
| 49 |
+
"experiment_timestamp": "2026-02-13T14:50:12.111719"
|
| 50 |
+
}
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json
ADDED
|
@@ -0,0 +1,578 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-13T14:46:01.975976",
|
| 4 |
+
"timestamp_ms": 1770961561975.988,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-13T14:46:31.976367",
|
| 10 |
+
"timestamp_ms": 1770961591976.404,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "skipped"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-13T14:46:31.976499",
|
| 16 |
+
"timestamp_ms": 1770961591976.5066,
|
| 17 |
+
"phase": "idle_mid",
|
| 18 |
+
"sub_phase": ""
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-13T14:47:01.976738",
|
| 22 |
+
"timestamp_ms": 1770961621976.7778,
|
| 23 |
+
"phase": "inference",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-13T14:47:05.466596",
|
| 28 |
+
"timestamp_ms": 1770961625466.6067,
|
| 29 |
+
"phase": "inference",
|
| 30 |
+
"sub_phase": "session_start_pattern_fixed"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-13T14:47:05.466707",
|
| 34 |
+
"timestamp_ms": 1770961625466.71,
|
| 35 |
+
"phase": "inference",
|
| 36 |
+
"sub_phase": "req_0_prefill"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-13T14:47:05.699872",
|
| 40 |
+
"timestamp_ms": 1770961625699.8838,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "req_0_decode"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-13T14:47:06.305844",
|
| 46 |
+
"timestamp_ms": 1770961626305.8555,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "req_0_idle_between"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-13T14:47:11.306143",
|
| 52 |
+
"timestamp_ms": 1770961631306.1824,
|
| 53 |
+
"phase": "inference",
|
| 54 |
+
"sub_phase": "req_1_prefill"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-13T14:47:11.313367",
|
| 58 |
+
"timestamp_ms": 1770961631313.379,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "req_1_decode"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-13T14:47:11.700522",
|
| 64 |
+
"timestamp_ms": 1770961631700.5303,
|
| 65 |
+
"phase": "inference",
|
| 66 |
+
"sub_phase": "req_1_idle_between"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-13T14:47:16.700777",
|
| 70 |
+
"timestamp_ms": 1770961636700.8154,
|
| 71 |
+
"phase": "inference",
|
| 72 |
+
"sub_phase": "req_2_prefill"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-13T14:47:16.707363",
|
| 76 |
+
"timestamp_ms": 1770961636707.374,
|
| 77 |
+
"phase": "inference",
|
| 78 |
+
"sub_phase": "req_2_decode"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-13T14:47:17.105804",
|
| 82 |
+
"timestamp_ms": 1770961637105.8113,
|
| 83 |
+
"phase": "inference",
|
| 84 |
+
"sub_phase": "req_2_idle_between"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-13T14:47:22.106068",
|
| 88 |
+
"timestamp_ms": 1770961642106.1057,
|
| 89 |
+
"phase": "inference",
|
| 90 |
+
"sub_phase": "req_3_prefill"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-13T14:47:22.115115",
|
| 94 |
+
"timestamp_ms": 1770961642115.1292,
|
| 95 |
+
"phase": "inference",
|
| 96 |
+
"sub_phase": "req_3_decode"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-13T14:47:22.512669",
|
| 100 |
+
"timestamp_ms": 1770961642512.6772,
|
| 101 |
+
"phase": "inference",
|
| 102 |
+
"sub_phase": "req_3_idle_between"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-13T14:47:27.512975",
|
| 106 |
+
"timestamp_ms": 1770961647513.0088,
|
| 107 |
+
"phase": "inference",
|
| 108 |
+
"sub_phase": "req_4_prefill"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-13T14:47:27.518782",
|
| 112 |
+
"timestamp_ms": 1770961647518.7922,
|
| 113 |
+
"phase": "inference",
|
| 114 |
+
"sub_phase": "req_4_decode"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-13T14:47:27.565196",
|
| 118 |
+
"timestamp_ms": 1770961647565.2043,
|
| 119 |
+
"phase": "inference",
|
| 120 |
+
"sub_phase": "req_4_idle_between"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-13T14:47:32.565447",
|
| 124 |
+
"timestamp_ms": 1770961652565.483,
|
| 125 |
+
"phase": "inference",
|
| 126 |
+
"sub_phase": "req_5_prefill"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-13T14:47:32.572933",
|
| 130 |
+
"timestamp_ms": 1770961652572.9426,
|
| 131 |
+
"phase": "inference",
|
| 132 |
+
"sub_phase": "req_5_decode"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-13T14:47:32.965121",
|
| 136 |
+
"timestamp_ms": 1770961652965.1296,
|
| 137 |
+
"phase": "inference",
|
| 138 |
+
"sub_phase": "req_5_idle_between"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-13T14:47:37.965395",
|
| 142 |
+
"timestamp_ms": 1770961657965.4297,
|
| 143 |
+
"phase": "inference",
|
| 144 |
+
"sub_phase": "req_6_prefill"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-13T14:47:38.012554",
|
| 148 |
+
"timestamp_ms": 1770961658012.5662,
|
| 149 |
+
"phase": "inference",
|
| 150 |
+
"sub_phase": "req_6_decode"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-13T14:47:38.420400",
|
| 154 |
+
"timestamp_ms": 1770961658420.4077,
|
| 155 |
+
"phase": "inference",
|
| 156 |
+
"sub_phase": "req_6_idle_between"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-13T14:47:43.420660",
|
| 160 |
+
"timestamp_ms": 1770961663420.6963,
|
| 161 |
+
"phase": "inference",
|
| 162 |
+
"sub_phase": "req_7_prefill"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-13T14:47:43.427244",
|
| 166 |
+
"timestamp_ms": 1770961663427.2568,
|
| 167 |
+
"phase": "inference",
|
| 168 |
+
"sub_phase": "req_7_decode"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-13T14:47:43.818844",
|
| 172 |
+
"timestamp_ms": 1770961663818.8525,
|
| 173 |
+
"phase": "inference",
|
| 174 |
+
"sub_phase": "req_7_idle_between"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-13T14:47:48.819102",
|
| 178 |
+
"timestamp_ms": 1770961668819.1387,
|
| 179 |
+
"phase": "inference",
|
| 180 |
+
"sub_phase": "req_8_prefill"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-13T14:47:48.825945",
|
| 184 |
+
"timestamp_ms": 1770961668825.9563,
|
| 185 |
+
"phase": "inference",
|
| 186 |
+
"sub_phase": "req_8_decode"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-13T14:47:49.223201",
|
| 190 |
+
"timestamp_ms": 1770961669223.2083,
|
| 191 |
+
"phase": "inference",
|
| 192 |
+
"sub_phase": "req_8_idle_between"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-13T14:47:54.223439",
|
| 196 |
+
"timestamp_ms": 1770961674223.475,
|
| 197 |
+
"phase": "inference",
|
| 198 |
+
"sub_phase": "req_9_prefill"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-13T14:47:54.230350",
|
| 202 |
+
"timestamp_ms": 1770961674230.3623,
|
| 203 |
+
"phase": "inference",
|
| 204 |
+
"sub_phase": "req_9_decode"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-13T14:47:54.620827",
|
| 208 |
+
"timestamp_ms": 1770961674620.843,
|
| 209 |
+
"phase": "inference",
|
| 210 |
+
"sub_phase": "req_9_idle_between"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-13T14:47:59.621088",
|
| 214 |
+
"timestamp_ms": 1770961679621.1243,
|
| 215 |
+
"phase": "inference",
|
| 216 |
+
"sub_phase": "req_10_prefill"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-13T14:47:59.627036",
|
| 220 |
+
"timestamp_ms": 1770961679627.0483,
|
| 221 |
+
"phase": "inference",
|
| 222 |
+
"sub_phase": "req_10_decode"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-13T14:48:00.022441",
|
| 226 |
+
"timestamp_ms": 1770961680022.4487,
|
| 227 |
+
"phase": "inference",
|
| 228 |
+
"sub_phase": "req_10_idle_between"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-13T14:48:05.022693",
|
| 232 |
+
"timestamp_ms": 1770961685022.7278,
|
| 233 |
+
"phase": "inference",
|
| 234 |
+
"sub_phase": "req_11_prefill"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-13T14:48:05.029345",
|
| 238 |
+
"timestamp_ms": 1770961685029.3545,
|
| 239 |
+
"phase": "inference",
|
| 240 |
+
"sub_phase": "req_11_decode"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-13T14:48:05.422593",
|
| 244 |
+
"timestamp_ms": 1770961685422.6013,
|
| 245 |
+
"phase": "inference",
|
| 246 |
+
"sub_phase": "req_11_idle_between"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-13T14:48:10.422898",
|
| 250 |
+
"timestamp_ms": 1770961690422.9382,
|
| 251 |
+
"phase": "inference",
|
| 252 |
+
"sub_phase": "req_12_prefill"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-13T14:48:10.430562",
|
| 256 |
+
"timestamp_ms": 1770961690430.5754,
|
| 257 |
+
"phase": "inference",
|
| 258 |
+
"sub_phase": "req_12_decode"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-13T14:48:10.831236",
|
| 262 |
+
"timestamp_ms": 1770961690831.2441,
|
| 263 |
+
"phase": "inference",
|
| 264 |
+
"sub_phase": "req_12_idle_between"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-13T14:48:15.831506",
|
| 268 |
+
"timestamp_ms": 1770961695831.5442,
|
| 269 |
+
"phase": "inference",
|
| 270 |
+
"sub_phase": "req_13_prefill"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-13T14:48:15.837978",
|
| 274 |
+
"timestamp_ms": 1770961695837.9888,
|
| 275 |
+
"phase": "inference",
|
| 276 |
+
"sub_phase": "req_13_decode"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-13T14:48:16.238816",
|
| 280 |
+
"timestamp_ms": 1770961696238.8247,
|
| 281 |
+
"phase": "inference",
|
| 282 |
+
"sub_phase": "req_13_idle_between"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-13T14:48:21.239113",
|
| 286 |
+
"timestamp_ms": 1770961701239.1526,
|
| 287 |
+
"phase": "inference",
|
| 288 |
+
"sub_phase": "req_14_prefill"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-13T14:48:21.245694",
|
| 292 |
+
"timestamp_ms": 1770961701245.7063,
|
| 293 |
+
"phase": "inference",
|
| 294 |
+
"sub_phase": "req_14_decode"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-13T14:48:21.638186",
|
| 298 |
+
"timestamp_ms": 1770961701638.195,
|
| 299 |
+
"phase": "inference",
|
| 300 |
+
"sub_phase": "req_14_idle_between"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-13T14:48:26.638431",
|
| 304 |
+
"timestamp_ms": 1770961706638.4685,
|
| 305 |
+
"phase": "inference",
|
| 306 |
+
"sub_phase": "req_15_prefill"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-13T14:48:26.645556",
|
| 310 |
+
"timestamp_ms": 1770961706645.5667,
|
| 311 |
+
"phase": "inference",
|
| 312 |
+
"sub_phase": "req_15_decode"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-13T14:48:27.045641",
|
| 316 |
+
"timestamp_ms": 1770961707045.6492,
|
| 317 |
+
"phase": "inference",
|
| 318 |
+
"sub_phase": "req_15_idle_between"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-13T14:48:32.045947",
|
| 322 |
+
"timestamp_ms": 1770961712045.9856,
|
| 323 |
+
"phase": "inference",
|
| 324 |
+
"sub_phase": "req_16_prefill"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-13T14:48:32.052755",
|
| 328 |
+
"timestamp_ms": 1770961712052.7668,
|
| 329 |
+
"phase": "inference",
|
| 330 |
+
"sub_phase": "req_16_decode"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-13T14:48:32.087561",
|
| 334 |
+
"timestamp_ms": 1770961712087.5703,
|
| 335 |
+
"phase": "inference",
|
| 336 |
+
"sub_phase": "req_16_idle_between"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-13T14:48:37.087889",
|
| 340 |
+
"timestamp_ms": 1770961717087.9277,
|
| 341 |
+
"phase": "inference",
|
| 342 |
+
"sub_phase": "req_17_prefill"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-13T14:48:37.094320",
|
| 346 |
+
"timestamp_ms": 1770961717094.333,
|
| 347 |
+
"phase": "inference",
|
| 348 |
+
"sub_phase": "req_17_decode"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-13T14:48:37.489046",
|
| 352 |
+
"timestamp_ms": 1770961717489.055,
|
| 353 |
+
"phase": "inference",
|
| 354 |
+
"sub_phase": "req_17_idle_between"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-13T14:48:42.489281",
|
| 358 |
+
"timestamp_ms": 1770961722489.3171,
|
| 359 |
+
"phase": "inference",
|
| 360 |
+
"sub_phase": "req_18_prefill"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-13T14:48:42.496940",
|
| 364 |
+
"timestamp_ms": 1770961722496.949,
|
| 365 |
+
"phase": "inference",
|
| 366 |
+
"sub_phase": "req_18_decode"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-13T14:48:42.889447",
|
| 370 |
+
"timestamp_ms": 1770961722889.4595,
|
| 371 |
+
"phase": "inference",
|
| 372 |
+
"sub_phase": "req_18_idle_between"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-13T14:48:47.889704",
|
| 376 |
+
"timestamp_ms": 1770961727889.7422,
|
| 377 |
+
"phase": "inference",
|
| 378 |
+
"sub_phase": "req_19_prefill"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-13T14:48:47.896785",
|
| 382 |
+
"timestamp_ms": 1770961727896.797,
|
| 383 |
+
"phase": "inference",
|
| 384 |
+
"sub_phase": "req_19_decode"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-13T14:48:48.292153",
|
| 388 |
+
"timestamp_ms": 1770961728292.161,
|
| 389 |
+
"phase": "inference",
|
| 390 |
+
"sub_phase": "req_19_idle_between"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-13T14:48:53.292443",
|
| 394 |
+
"timestamp_ms": 1770961733292.4824,
|
| 395 |
+
"phase": "inference",
|
| 396 |
+
"sub_phase": "req_20_prefill"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-13T14:48:53.298830",
|
| 400 |
+
"timestamp_ms": 1770961733298.848,
|
| 401 |
+
"phase": "inference",
|
| 402 |
+
"sub_phase": "req_20_decode"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-13T14:48:53.698341",
|
| 406 |
+
"timestamp_ms": 1770961733698.349,
|
| 407 |
+
"phase": "inference",
|
| 408 |
+
"sub_phase": "req_20_idle_between"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-13T14:48:58.698584",
|
| 412 |
+
"timestamp_ms": 1770961738698.6228,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "req_21_prefill"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-13T14:48:58.704854",
|
| 418 |
+
"timestamp_ms": 1770961738704.8652,
|
| 419 |
+
"phase": "inference",
|
| 420 |
+
"sub_phase": "req_21_decode"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-13T14:48:59.095313",
|
| 424 |
+
"timestamp_ms": 1770961739095.3215,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "req_21_idle_between"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-13T14:49:04.095567",
|
| 430 |
+
"timestamp_ms": 1770961744095.6064,
|
| 431 |
+
"phase": "inference",
|
| 432 |
+
"sub_phase": "req_22_prefill"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-13T14:49:04.101884",
|
| 436 |
+
"timestamp_ms": 1770961744101.8953,
|
| 437 |
+
"phase": "inference",
|
| 438 |
+
"sub_phase": "req_22_decode"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-13T14:49:04.496594",
|
| 442 |
+
"timestamp_ms": 1770961744496.6016,
|
| 443 |
+
"phase": "inference",
|
| 444 |
+
"sub_phase": "req_22_idle_between"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-13T14:49:09.496919",
|
| 448 |
+
"timestamp_ms": 1770961749496.959,
|
| 449 |
+
"phase": "inference",
|
| 450 |
+
"sub_phase": "req_23_prefill"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-13T14:49:09.506605",
|
| 454 |
+
"timestamp_ms": 1770961749506.618,
|
| 455 |
+
"phase": "inference",
|
| 456 |
+
"sub_phase": "req_23_decode"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-13T14:49:09.714774",
|
| 460 |
+
"timestamp_ms": 1770961749714.782,
|
| 461 |
+
"phase": "inference",
|
| 462 |
+
"sub_phase": "req_23_idle_between"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-13T14:49:14.715088",
|
| 466 |
+
"timestamp_ms": 1770961754715.1287,
|
| 467 |
+
"phase": "inference",
|
| 468 |
+
"sub_phase": "req_24_prefill"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-13T14:49:14.721983",
|
| 472 |
+
"timestamp_ms": 1770961754721.9963,
|
| 473 |
+
"phase": "inference",
|
| 474 |
+
"sub_phase": "req_24_decode"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-13T14:49:15.114449",
|
| 478 |
+
"timestamp_ms": 1770961755114.457,
|
| 479 |
+
"phase": "inference",
|
| 480 |
+
"sub_phase": "req_24_idle_between"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-13T14:49:20.114694",
|
| 484 |
+
"timestamp_ms": 1770961760114.73,
|
| 485 |
+
"phase": "inference",
|
| 486 |
+
"sub_phase": "req_25_prefill"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-13T14:49:20.121468",
|
| 490 |
+
"timestamp_ms": 1770961760121.4773,
|
| 491 |
+
"phase": "inference",
|
| 492 |
+
"sub_phase": "req_25_decode"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-13T14:49:20.509052",
|
| 496 |
+
"timestamp_ms": 1770961760509.0598,
|
| 497 |
+
"phase": "inference",
|
| 498 |
+
"sub_phase": "req_25_idle_between"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-13T14:49:25.509322",
|
| 502 |
+
"timestamp_ms": 1770961765509.3591,
|
| 503 |
+
"phase": "inference",
|
| 504 |
+
"sub_phase": "req_26_prefill"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-13T14:49:25.515483",
|
| 508 |
+
"timestamp_ms": 1770961765515.4946,
|
| 509 |
+
"phase": "inference",
|
| 510 |
+
"sub_phase": "req_26_decode"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-13T14:49:25.903568",
|
| 514 |
+
"timestamp_ms": 1770961765903.5762,
|
| 515 |
+
"phase": "inference",
|
| 516 |
+
"sub_phase": "req_26_idle_between"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-13T14:49:30.903929",
|
| 520 |
+
"timestamp_ms": 1770961770903.9673,
|
| 521 |
+
"phase": "inference",
|
| 522 |
+
"sub_phase": "req_27_prefill"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-13T14:49:30.910400",
|
| 526 |
+
"timestamp_ms": 1770961770910.4119,
|
| 527 |
+
"phase": "inference",
|
| 528 |
+
"sub_phase": "req_27_decode"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-13T14:49:31.307382",
|
| 532 |
+
"timestamp_ms": 1770961771307.3904,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "req_27_idle_between"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-13T14:49:36.307700",
|
| 538 |
+
"timestamp_ms": 1770961776307.739,
|
| 539 |
+
"phase": "inference",
|
| 540 |
+
"sub_phase": "req_28_prefill"
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"timestamp": "2026-02-13T14:49:36.313937",
|
| 544 |
+
"timestamp_ms": 1770961776313.9487,
|
| 545 |
+
"phase": "inference",
|
| 546 |
+
"sub_phase": "req_28_decode"
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"timestamp": "2026-02-13T14:49:36.703704",
|
| 550 |
+
"timestamp_ms": 1770961776703.7122,
|
| 551 |
+
"phase": "inference",
|
| 552 |
+
"sub_phase": "req_28_idle_between"
|
| 553 |
+
},
|
| 554 |
+
{
|
| 555 |
+
"timestamp": "2026-02-13T14:49:41.704041",
|
| 556 |
+
"timestamp_ms": 1770961781704.0808,
|
| 557 |
+
"phase": "inference",
|
| 558 |
+
"sub_phase": "req_29_prefill"
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"timestamp": "2026-02-13T14:49:41.711155",
|
| 562 |
+
"timestamp_ms": 1770961781711.168,
|
| 563 |
+
"phase": "inference",
|
| 564 |
+
"sub_phase": "req_29_decode"
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"timestamp": "2026-02-13T14:49:42.099889",
|
| 568 |
+
"timestamp_ms": 1770961782099.898,
|
| 569 |
+
"phase": "inference",
|
| 570 |
+
"sub_phase": "session_finished"
|
| 571 |
+
},
|
| 572 |
+
{
|
| 573 |
+
"timestamp": "2026-02-13T14:49:42.099988",
|
| 574 |
+
"timestamp_ms": 1770961782099.9902,
|
| 575 |
+
"phase": "idle_post",
|
| 576 |
+
"sub_phase": ""
|
| 577 |
+
}
|
| 578 |
+
]
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
request_idx,input_tokens,output_tokens,prefill_time_ms,decode_time_ms,prompt_preview,output_preview
|
| 2 |
+
0,7,128,214.58721160888672,605.3135395050049,Give three tips for staying healthy.,"
|
| 3 |
+
|
| 4 |
+
1. Buy healthy food.
|
| 5 |
+
|
| 6 |
+
Some people are so desperate to get healthy they want to"
|
| 7 |
+
1,8,128,5.532264709472656,386.86227798461914,Design an app for a delivery company.,"
|
| 8 |
+
|
| 9 |
+
This is a small app that lets you do all sorts of things.
|
| 10 |
+
|
| 11 |
+
This app is for eve"
|
| 12 |
+
2,11,128,4.894733428955078,398.1354236602783,Describe the life and reign of King Charles II.,"
|
| 13 |
+
|
| 14 |
+
The story begins with a visit to the Royal Children's Hospital in London in 16"
|
| 15 |
+
3,8,128,7.489681243896484,397.2821235656738,Render a 3D model of a house," from a new angle, showing the same angle in all three dimensions. The house is "
|
| 16 |
+
4,11,2,4.428386688232422,11.75379753112793,Propose an ethical solution to the problem of data privacy,.
|
| 17 |
+
5,53,128,5.54656982421875,391.9224739074707,"Reverse engineer this code to create a new version
|
| 18 |
+
def factorialize(num):
|
| 19 |
+
fact","
|
| 20 |
+
def doxygen(num): for i in range(1, num):
|
| 21 |
+
for num in count(num): "
|
| 22 |
+
6,20,128,5.455970764160156,407.58371353149414,Write a short story in third person narration about a protagonist who has to mak,"
|
| 23 |
+
|
| 24 |
+
Narrate your story directly to the reader. You can add the words of your hero'"
|
| 25 |
+
7,6,128,5.154132843017578,391.31665229797363,How did Julius Caesar die?, Is it possible that the Roman Emperor died during the reign of Augustus? For ex
|
| 26 |
+
8,25,128,5.235433578491211,396.99387550354004,"Given a set of numbers, find the maximum value.
|
| 27 |
+
Set: {10, 3, 25, 6, 16}","
|
| 28 |
+
Set of numbers: {0, 2, 6, 5, 4}
|
| 29 |
+
Set of values: {0, 1, 10}
|
| 30 |
+
Set of values: {0, 3,"
|
| 31 |
+
9,23,128,5.2738189697265625,390.2015686035156,Arrange the words in the given sentence to form a grammatically correct sentence," out of his nest and started running towards the door.
|
| 32 |
+
""I'll take you out when y"
|
| 33 |
+
10,19,128,4.380702972412109,395.11728286743164,"Explain why the given definition is wrong.
|
| 34 |
+
A mole is an animal that lives underg", Its food is usually a mix of soil and water. The mole has a body mass of around
|
| 35 |
+
11,19,128,5.073308944702148,392.9727077484131,"Convert the given equation into an algebraic expression.
|
| 36 |
+
3x+5y=9","x/10y=13x/14x=20x/25y=35x/40y+60y=80x/90y+110x/120y
|
| 37 |
+
The above equation can be do"
|
| 38 |
+
12,18,128,5.993843078613281,400.38371086120605,Calculate the total surface area of a cube with a side length of 5 cm.,"
|
| 39 |
+
|
| 40 |
+
To find the minimum depth of the cube, divide the distance between the two edg"
|
| 41 |
+
13,9,128,4.939794540405273,400.5284309387207,Design a logo for a website about recycling.,"
|
| 42 |
+
|
| 43 |
+
The idea is that after the site has been redesigned, with the site being the f"
|
| 44 |
+
14,9,128,5.007743835449219,392.1947479248047,Explain the concept of cogging torque.,"
|
| 45 |
+
|
| 46 |
+
Let's say you're in a race. You want to know how many steps you need to make t"
|
| 47 |
+
15,11,128,5.470514297485352,399.7819423675537,Explain the process of cellular respiration in plants.,"
|
| 48 |
+
|
| 49 |
+
Proliferative Chemistry
|
| 50 |
+
|
| 51 |
+
The chemists have identified a chemical product of th"
|
| 52 |
+
16,7,9,5.344390869140625,34.53683853149414,Discuss the causes of the Great Depression, and what it means for people today.
|
| 53 |
+
17,15,128,4.845857620239258,394.4530487060547,What is the force on a 1 kg mass due to the gravitational force?,"
|
| 54 |
+
|
| 55 |
+
The force on a 1 kg mass is the force divided by the mass divided by the mass."
|
| 56 |
+
18,13,128,6.186485290527344,392.0886516571045,Describe a time when you had to make a difficult decision.,"
|
| 57 |
+
|
| 58 |
+
Your decision to commit suicide by suicide was painful. But now you are experi"
|
| 59 |
+
19,16,128,5.428314208984375,395.0660228729248,"Explain why the following fraction is equivalent to 1/4
|
| 60 |
+
4/16"," = 1.75 = 14.25
|
| 61 |
+
5/16 = 15.25 = 16.25 = 31.25
|
| 62 |
+
6/16 = 19.25 = 24.25 = 38.25
|
| 63 |
+
7/16 ="
|
| 64 |
+
20,15,128,4.713296890258789,399.2118835449219,Generate a list of ten items a person might need for a camping trip,", and then choose which one to use. You can choose the value of the first item i"
|
| 65 |
+
21,8,128,4.785776138305664,390.1824951171875,Describe the structure of an atom.,"
|
| 66 |
+
|
| 67 |
+
The atom consists of two parts: an atom that has a nucleus and a spin. A spin "
|
| 68 |
+
22,7,128,4.740238189697266,394.42896842956543,What are the three primary colors?,"
|
| 69 |
+
|
| 70 |
+
Pink, Blue, and Yellow.
|
| 71 |
+
|
| 72 |
+
What color will be used in the game?
|
| 73 |
+
|
| 74 |
+
The following t"
|
| 75 |
+
23,21,66,8.002042770385742,207.89098739624023,"Use the given data to calculate the median.
|
| 76 |
+
[2, 3, 7, 8, 10]","
|
| 77 |
+
There are three ways to calculate the median.
|
| 78 |
+
[1, 2, 3, 4, 5, 6, 11]
|
| 79 |
+
[2, 3, 4, "
|
| 80 |
+
24,9,128,5.3234100341796875,392.17686653137207,Who is the world's most famous painter?, What are the most famous paintings? Who is the world's most famous musician? Wh
|
| 81 |
+
25,7,128,5.42759895324707,387.3109817504883,How can we reduce air pollution?,"
|
| 82 |
+
|
| 83 |
+
The answer to this question is complicated and we don't want to just jump to c"
|
| 84 |
+
26,23,128,4.314661026000977,387.8059387207031,"Evaluate this sentence for spelling and grammar mistakes
|
| 85 |
+
He finnished his meal a"," alone.
|
| 86 |
+
'She's not going to be here before she's gone,' he said.
|
| 87 |
+
He glanced over"
|
| 88 |
+
27,19,128,4.817485809326172,396.6984748840332,"Rewrite the following sentence using active voice.
|
| 89 |
+
The news report was read by t","
|
| 90 |
+
The captain said something to the effect of ""I will be the first one to go to M"
|
| 91 |
+
28,13,128,4.557371139526367,389.47558403015137,Explain the use of word embeddings in Natural Language Processing," (NLP).
|
| 92 |
+
|
| 93 |
+
The second example illustrates the use of word embeddings in Natural La"
|
| 94 |
+
29,7,128,5.580425262451172,388.4294033050537,Outline the consequences of deforestation.,"
|
| 95 |
+
|
| 96 |
+
In many places, the amount of water the Amazon consumes is much higher than th"
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_gpt2_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3335,33.483248,309.625786,9.250325,14.635,14.765,7.574
|
| 3 |
+
idle_post,3000,29.989958,1608.086256,53.605432,78.325,91.001,6.969
|
| 4 |
+
idle_pre,3001,30.01006,1612.965583,53.747406,78.435,79.647,7.522
|
| 5 |
+
inference_decode,1114,11.139613,852.811078,76.556568,86.774,91.001,8.911
|
| 6 |
+
inference_idle,14501,145.01153,11922.965078,82.220878,96.495,102.422,43.361
|
| 7 |
+
inference_prefill,49,0.488904,21.451692,43.959224,78.2042,78.405,7.783
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 3,
|
| 7 |
+
"train_max_steps": null,
|
| 8 |
+
"batch_size": 4,
|
| 9 |
+
"gradient_accumulation_steps": 4,
|
| 10 |
+
"learning_rate": 0.0002,
|
| 11 |
+
"lora_r": 16,
|
| 12 |
+
"lora_alpha": 32,
|
| 13 |
+
"lora_dropout": 0.05,
|
| 14 |
+
"max_seq_length": 512,
|
| 15 |
+
"use_4bit": false,
|
| 16 |
+
"use_bf16": true,
|
| 17 |
+
"inference_total_requests": 30,
|
| 18 |
+
"inference_interval_sec": 5.0,
|
| 19 |
+
"inference_pattern": "fixed",
|
| 20 |
+
"inference_variable_min": 2.0,
|
| 21 |
+
"inference_variable_max": 10.0,
|
| 22 |
+
"inference_burst_count": 3,
|
| 23 |
+
"inference_burst_gap": 1.0,
|
| 24 |
+
"inference_burst_pause": 15.0,
|
| 25 |
+
"gen_max_new_tokens": 128,
|
| 26 |
+
"gen_temperature": 0.7,
|
| 27 |
+
"gen_do_sample": true,
|
| 28 |
+
"prompt_set": "alpaca",
|
| 29 |
+
"clock_lock_gpu": null,
|
| 30 |
+
"clock_lock_mem": null,
|
| 31 |
+
"ramp_enabled": false,
|
| 32 |
+
"ramp_start_w": null,
|
| 33 |
+
"ramp_end_w": null,
|
| 34 |
+
"ramp_steps": 5,
|
| 35 |
+
"ramp_step_duration": 3.0,
|
| 36 |
+
"device": "cuda:0",
|
| 37 |
+
"gpu_index": 0,
|
| 38 |
+
"skip_training": true,
|
| 39 |
+
"skip_inference": false,
|
| 40 |
+
"model": "llama3.1-8b",
|
| 41 |
+
"power_cap": null,
|
| 42 |
+
"gpu_info": {
|
| 43 |
+
"gpu_name": "NVIDIA RTX PRO 6000 Blackwell Workstation Edition",
|
| 44 |
+
"tdp_w": 600,
|
| 45 |
+
"memory_total_mb": 97887,
|
| 46 |
+
"driver_version": "580.82.07",
|
| 47 |
+
"cuda_version": "13.0"
|
| 48 |
+
},
|
| 49 |
+
"experiment_timestamp": "2026-02-13T15:16:27.375097"
|
| 50 |
+
}
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json
ADDED
|
@@ -0,0 +1,578 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-13T15:11:35.115114",
|
| 4 |
+
"timestamp_ms": 1770963095115.1238,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-13T15:12:05.115548",
|
| 10 |
+
"timestamp_ms": 1770963125115.5862,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "skipped"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-13T15:12:05.115682",
|
| 16 |
+
"timestamp_ms": 1770963125115.6904,
|
| 17 |
+
"phase": "idle_mid",
|
| 18 |
+
"sub_phase": ""
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-13T15:12:35.115951",
|
| 22 |
+
"timestamp_ms": 1770963155115.991,
|
| 23 |
+
"phase": "inference",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-13T15:12:38.476499",
|
| 28 |
+
"timestamp_ms": 1770963158476.5122,
|
| 29 |
+
"phase": "inference",
|
| 30 |
+
"sub_phase": "session_start_pattern_fixed"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-13T15:12:38.476628",
|
| 34 |
+
"timestamp_ms": 1770963158476.6309,
|
| 35 |
+
"phase": "inference",
|
| 36 |
+
"sub_phase": "req_0_prefill"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-13T15:12:38.816631",
|
| 40 |
+
"timestamp_ms": 1770963158816.6428,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "req_0_decode"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-13T15:12:40.792783",
|
| 46 |
+
"timestamp_ms": 1770963160792.813,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "req_0_idle_between"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-13T15:12:45.793181",
|
| 52 |
+
"timestamp_ms": 1770963165793.217,
|
| 53 |
+
"phase": "inference",
|
| 54 |
+
"sub_phase": "req_1_prefill"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-13T15:12:45.810557",
|
| 58 |
+
"timestamp_ms": 1770963165810.568,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "req_1_decode"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-13T15:12:47.568807",
|
| 64 |
+
"timestamp_ms": 1770963167568.8147,
|
| 65 |
+
"phase": "inference",
|
| 66 |
+
"sub_phase": "req_1_idle_between"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-13T15:12:52.569070",
|
| 70 |
+
"timestamp_ms": 1770963172569.1052,
|
| 71 |
+
"phase": "inference",
|
| 72 |
+
"sub_phase": "req_2_prefill"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-13T15:12:52.586916",
|
| 76 |
+
"timestamp_ms": 1770963172586.926,
|
| 77 |
+
"phase": "inference",
|
| 78 |
+
"sub_phase": "req_2_decode"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-13T15:12:54.345393",
|
| 82 |
+
"timestamp_ms": 1770963174345.4011,
|
| 83 |
+
"phase": "inference",
|
| 84 |
+
"sub_phase": "req_2_idle_between"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-13T15:12:59.345647",
|
| 88 |
+
"timestamp_ms": 1770963179345.6824,
|
| 89 |
+
"phase": "inference",
|
| 90 |
+
"sub_phase": "req_3_prefill"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-13T15:12:59.364781",
|
| 94 |
+
"timestamp_ms": 1770963179364.792,
|
| 95 |
+
"phase": "inference",
|
| 96 |
+
"sub_phase": "req_3_decode"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-13T15:13:01.126247",
|
| 100 |
+
"timestamp_ms": 1770963181126.2556,
|
| 101 |
+
"phase": "inference",
|
| 102 |
+
"sub_phase": "req_3_idle_between"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-13T15:13:06.126504",
|
| 106 |
+
"timestamp_ms": 1770963186126.539,
|
| 107 |
+
"phase": "inference",
|
| 108 |
+
"sub_phase": "req_4_prefill"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-13T15:13:06.144131",
|
| 112 |
+
"timestamp_ms": 1770963186144.142,
|
| 113 |
+
"phase": "inference",
|
| 114 |
+
"sub_phase": "req_4_decode"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-13T15:13:07.903177",
|
| 118 |
+
"timestamp_ms": 1770963187903.185,
|
| 119 |
+
"phase": "inference",
|
| 120 |
+
"sub_phase": "req_4_idle_between"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-13T15:13:12.903415",
|
| 124 |
+
"timestamp_ms": 1770963192903.4475,
|
| 125 |
+
"phase": "inference",
|
| 126 |
+
"sub_phase": "req_5_prefill"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-13T15:13:12.921571",
|
| 130 |
+
"timestamp_ms": 1770963192921.5798,
|
| 131 |
+
"phase": "inference",
|
| 132 |
+
"sub_phase": "req_5_decode"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-13T15:13:14.682712",
|
| 136 |
+
"timestamp_ms": 1770963194682.72,
|
| 137 |
+
"phase": "inference",
|
| 138 |
+
"sub_phase": "req_5_idle_between"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-13T15:13:19.682976",
|
| 142 |
+
"timestamp_ms": 1770963199683.0093,
|
| 143 |
+
"phase": "inference",
|
| 144 |
+
"sub_phase": "req_6_prefill"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-13T15:13:19.701522",
|
| 148 |
+
"timestamp_ms": 1770963199701.5317,
|
| 149 |
+
"phase": "inference",
|
| 150 |
+
"sub_phase": "req_6_decode"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-13T15:13:21.461635",
|
| 154 |
+
"timestamp_ms": 1770963201461.6428,
|
| 155 |
+
"phase": "inference",
|
| 156 |
+
"sub_phase": "req_6_idle_between"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-13T15:13:26.461931",
|
| 160 |
+
"timestamp_ms": 1770963206461.9663,
|
| 161 |
+
"phase": "inference",
|
| 162 |
+
"sub_phase": "req_7_prefill"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-13T15:13:26.479652",
|
| 166 |
+
"timestamp_ms": 1770963206479.6636,
|
| 167 |
+
"phase": "inference",
|
| 168 |
+
"sub_phase": "req_7_decode"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-13T15:13:28.239291",
|
| 172 |
+
"timestamp_ms": 1770963208239.2996,
|
| 173 |
+
"phase": "inference",
|
| 174 |
+
"sub_phase": "req_7_idle_between"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-13T15:13:33.239587",
|
| 178 |
+
"timestamp_ms": 1770963213239.6216,
|
| 179 |
+
"phase": "inference",
|
| 180 |
+
"sub_phase": "req_8_prefill"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-13T15:13:33.258046",
|
| 184 |
+
"timestamp_ms": 1770963213258.058,
|
| 185 |
+
"phase": "inference",
|
| 186 |
+
"sub_phase": "req_8_decode"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-13T15:13:35.018557",
|
| 190 |
+
"timestamp_ms": 1770963215018.566,
|
| 191 |
+
"phase": "inference",
|
| 192 |
+
"sub_phase": "req_8_idle_between"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-13T15:13:40.018802",
|
| 196 |
+
"timestamp_ms": 1770963220018.876,
|
| 197 |
+
"phase": "inference",
|
| 198 |
+
"sub_phase": "req_9_prefill"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-13T15:13:40.035668",
|
| 202 |
+
"timestamp_ms": 1770963220035.6794,
|
| 203 |
+
"phase": "inference",
|
| 204 |
+
"sub_phase": "req_9_decode"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-13T15:13:41.796109",
|
| 208 |
+
"timestamp_ms": 1770963221796.117,
|
| 209 |
+
"phase": "inference",
|
| 210 |
+
"sub_phase": "req_9_idle_between"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-13T15:13:46.796271",
|
| 214 |
+
"timestamp_ms": 1770963226796.3027,
|
| 215 |
+
"phase": "inference",
|
| 216 |
+
"sub_phase": "req_10_prefill"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-13T15:13:46.814891",
|
| 220 |
+
"timestamp_ms": 1770963226814.902,
|
| 221 |
+
"phase": "inference",
|
| 222 |
+
"sub_phase": "req_10_decode"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-13T15:13:48.576538",
|
| 226 |
+
"timestamp_ms": 1770963228576.5452,
|
| 227 |
+
"phase": "inference",
|
| 228 |
+
"sub_phase": "req_10_idle_between"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-13T15:13:53.576731",
|
| 232 |
+
"timestamp_ms": 1770963233576.764,
|
| 233 |
+
"phase": "inference",
|
| 234 |
+
"sub_phase": "req_11_prefill"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-13T15:13:53.595518",
|
| 238 |
+
"timestamp_ms": 1770963233595.5295,
|
| 239 |
+
"phase": "inference",
|
| 240 |
+
"sub_phase": "req_11_decode"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-13T15:13:55.354287",
|
| 244 |
+
"timestamp_ms": 1770963235354.295,
|
| 245 |
+
"phase": "inference",
|
| 246 |
+
"sub_phase": "req_11_idle_between"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-13T15:14:00.354516",
|
| 250 |
+
"timestamp_ms": 1770963240354.5515,
|
| 251 |
+
"phase": "inference",
|
| 252 |
+
"sub_phase": "req_12_prefill"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-13T15:14:00.371439",
|
| 256 |
+
"timestamp_ms": 1770963240371.4492,
|
| 257 |
+
"phase": "inference",
|
| 258 |
+
"sub_phase": "req_12_decode"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-13T15:14:02.127329",
|
| 262 |
+
"timestamp_ms": 1770963242127.337,
|
| 263 |
+
"phase": "inference",
|
| 264 |
+
"sub_phase": "req_12_idle_between"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-13T15:14:07.127536",
|
| 268 |
+
"timestamp_ms": 1770963247127.5698,
|
| 269 |
+
"phase": "inference",
|
| 270 |
+
"sub_phase": "req_13_prefill"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-13T15:14:07.145196",
|
| 274 |
+
"timestamp_ms": 1770963247145.2065,
|
| 275 |
+
"phase": "inference",
|
| 276 |
+
"sub_phase": "req_13_decode"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-13T15:14:08.901022",
|
| 280 |
+
"timestamp_ms": 1770963248901.0295,
|
| 281 |
+
"phase": "inference",
|
| 282 |
+
"sub_phase": "req_13_idle_between"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-13T15:14:13.901235",
|
| 286 |
+
"timestamp_ms": 1770963253901.269,
|
| 287 |
+
"phase": "inference",
|
| 288 |
+
"sub_phase": "req_14_prefill"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-13T15:14:13.918636",
|
| 292 |
+
"timestamp_ms": 1770963253918.6465,
|
| 293 |
+
"phase": "inference",
|
| 294 |
+
"sub_phase": "req_14_decode"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-13T15:14:15.676535",
|
| 298 |
+
"timestamp_ms": 1770963255676.5425,
|
| 299 |
+
"phase": "inference",
|
| 300 |
+
"sub_phase": "req_14_idle_between"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-13T15:14:20.676751",
|
| 304 |
+
"timestamp_ms": 1770963260676.7832,
|
| 305 |
+
"phase": "inference",
|
| 306 |
+
"sub_phase": "req_15_prefill"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-13T15:14:20.694143",
|
| 310 |
+
"timestamp_ms": 1770963260694.1533,
|
| 311 |
+
"phase": "inference",
|
| 312 |
+
"sub_phase": "req_15_decode"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-13T15:14:22.450344",
|
| 316 |
+
"timestamp_ms": 1770963262450.3523,
|
| 317 |
+
"phase": "inference",
|
| 318 |
+
"sub_phase": "req_15_idle_between"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-13T15:14:27.450587",
|
| 322 |
+
"timestamp_ms": 1770963267450.6204,
|
| 323 |
+
"phase": "inference",
|
| 324 |
+
"sub_phase": "req_16_prefill"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-13T15:14:27.468077",
|
| 328 |
+
"timestamp_ms": 1770963267468.0864,
|
| 329 |
+
"phase": "inference",
|
| 330 |
+
"sub_phase": "req_16_decode"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-13T15:14:29.225874",
|
| 334 |
+
"timestamp_ms": 1770963269225.8818,
|
| 335 |
+
"phase": "inference",
|
| 336 |
+
"sub_phase": "req_16_idle_between"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-13T15:14:34.226126",
|
| 340 |
+
"timestamp_ms": 1770963274226.1604,
|
| 341 |
+
"phase": "inference",
|
| 342 |
+
"sub_phase": "req_17_prefill"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-13T15:14:34.243522",
|
| 346 |
+
"timestamp_ms": 1770963274243.5334,
|
| 347 |
+
"phase": "inference",
|
| 348 |
+
"sub_phase": "req_17_decode"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-13T15:14:36.001711",
|
| 352 |
+
"timestamp_ms": 1770963276001.7192,
|
| 353 |
+
"phase": "inference",
|
| 354 |
+
"sub_phase": "req_17_idle_between"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-13T15:14:41.001998",
|
| 358 |
+
"timestamp_ms": 1770963281002.0315,
|
| 359 |
+
"phase": "inference",
|
| 360 |
+
"sub_phase": "req_18_prefill"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-13T15:14:41.042742",
|
| 364 |
+
"timestamp_ms": 1770963281042.753,
|
| 365 |
+
"phase": "inference",
|
| 366 |
+
"sub_phase": "req_18_decode"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-13T15:14:42.808161",
|
| 370 |
+
"timestamp_ms": 1770963282808.17,
|
| 371 |
+
"phase": "inference",
|
| 372 |
+
"sub_phase": "req_18_idle_between"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-13T15:14:47.808393",
|
| 376 |
+
"timestamp_ms": 1770963287808.4258,
|
| 377 |
+
"phase": "inference",
|
| 378 |
+
"sub_phase": "req_19_prefill"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-13T15:14:47.825383",
|
| 382 |
+
"timestamp_ms": 1770963287825.3918,
|
| 383 |
+
"phase": "inference",
|
| 384 |
+
"sub_phase": "req_19_decode"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-13T15:14:49.582972",
|
| 388 |
+
"timestamp_ms": 1770963289582.9802,
|
| 389 |
+
"phase": "inference",
|
| 390 |
+
"sub_phase": "req_19_idle_between"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-13T15:14:54.583193",
|
| 394 |
+
"timestamp_ms": 1770963294583.2288,
|
| 395 |
+
"phase": "inference",
|
| 396 |
+
"sub_phase": "req_20_prefill"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-13T15:14:54.601092",
|
| 400 |
+
"timestamp_ms": 1770963294601.1018,
|
| 401 |
+
"phase": "inference",
|
| 402 |
+
"sub_phase": "req_20_decode"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-13T15:14:56.358875",
|
| 406 |
+
"timestamp_ms": 1770963296358.8838,
|
| 407 |
+
"phase": "inference",
|
| 408 |
+
"sub_phase": "req_20_idle_between"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-13T15:15:01.359126",
|
| 412 |
+
"timestamp_ms": 1770963301359.1624,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "req_21_prefill"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-13T15:15:01.376481",
|
| 418 |
+
"timestamp_ms": 1770963301376.4905,
|
| 419 |
+
"phase": "inference",
|
| 420 |
+
"sub_phase": "req_21_decode"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-13T15:15:03.135814",
|
| 424 |
+
"timestamp_ms": 1770963303135.822,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "req_21_idle_between"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-13T15:15:08.136065",
|
| 430 |
+
"timestamp_ms": 1770963308136.1006,
|
| 431 |
+
"phase": "inference",
|
| 432 |
+
"sub_phase": "req_22_prefill"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-13T15:15:08.152977",
|
| 436 |
+
"timestamp_ms": 1770963308152.988,
|
| 437 |
+
"phase": "inference",
|
| 438 |
+
"sub_phase": "req_22_decode"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-13T15:15:09.913207",
|
| 442 |
+
"timestamp_ms": 1770963309913.2153,
|
| 443 |
+
"phase": "inference",
|
| 444 |
+
"sub_phase": "req_22_idle_between"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-13T15:15:14.913421",
|
| 448 |
+
"timestamp_ms": 1770963314913.4568,
|
| 449 |
+
"phase": "inference",
|
| 450 |
+
"sub_phase": "req_23_prefill"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-13T15:15:14.931497",
|
| 454 |
+
"timestamp_ms": 1770963314931.508,
|
| 455 |
+
"phase": "inference",
|
| 456 |
+
"sub_phase": "req_23_decode"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-13T15:15:16.693296",
|
| 460 |
+
"timestamp_ms": 1770963316693.3047,
|
| 461 |
+
"phase": "inference",
|
| 462 |
+
"sub_phase": "req_23_idle_between"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-13T15:15:21.693545",
|
| 466 |
+
"timestamp_ms": 1770963321693.5803,
|
| 467 |
+
"phase": "inference",
|
| 468 |
+
"sub_phase": "req_24_prefill"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-13T15:15:21.711021",
|
| 472 |
+
"timestamp_ms": 1770963321711.0315,
|
| 473 |
+
"phase": "inference",
|
| 474 |
+
"sub_phase": "req_24_decode"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-13T15:15:23.470324",
|
| 478 |
+
"timestamp_ms": 1770963323470.332,
|
| 479 |
+
"phase": "inference",
|
| 480 |
+
"sub_phase": "req_24_idle_between"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-13T15:15:28.470580",
|
| 484 |
+
"timestamp_ms": 1770963328470.6167,
|
| 485 |
+
"phase": "inference",
|
| 486 |
+
"sub_phase": "req_25_prefill"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-13T15:15:28.487759",
|
| 490 |
+
"timestamp_ms": 1770963328487.77,
|
| 491 |
+
"phase": "inference",
|
| 492 |
+
"sub_phase": "req_25_decode"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-13T15:15:30.245623",
|
| 496 |
+
"timestamp_ms": 1770963330245.6309,
|
| 497 |
+
"phase": "inference",
|
| 498 |
+
"sub_phase": "req_25_idle_between"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-13T15:15:35.245825",
|
| 502 |
+
"timestamp_ms": 1770963335245.9304,
|
| 503 |
+
"phase": "inference",
|
| 504 |
+
"sub_phase": "req_26_prefill"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-13T15:15:35.263742",
|
| 508 |
+
"timestamp_ms": 1770963335263.7532,
|
| 509 |
+
"phase": "inference",
|
| 510 |
+
"sub_phase": "req_26_decode"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-13T15:15:37.024545",
|
| 514 |
+
"timestamp_ms": 1770963337024.5535,
|
| 515 |
+
"phase": "inference",
|
| 516 |
+
"sub_phase": "req_26_idle_between"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-13T15:15:42.024770",
|
| 520 |
+
"timestamp_ms": 1770963342024.8057,
|
| 521 |
+
"phase": "inference",
|
| 522 |
+
"sub_phase": "req_27_prefill"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-13T15:15:42.041986",
|
| 526 |
+
"timestamp_ms": 1770963342041.9976,
|
| 527 |
+
"phase": "inference",
|
| 528 |
+
"sub_phase": "req_27_decode"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-13T15:15:43.801957",
|
| 532 |
+
"timestamp_ms": 1770963343801.9658,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "req_27_idle_between"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-13T15:15:48.802207",
|
| 538 |
+
"timestamp_ms": 1770963348802.2417,
|
| 539 |
+
"phase": "inference",
|
| 540 |
+
"sub_phase": "req_28_prefill"
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"timestamp": "2026-02-13T15:15:48.820129",
|
| 544 |
+
"timestamp_ms": 1770963348820.139,
|
| 545 |
+
"phase": "inference",
|
| 546 |
+
"sub_phase": "req_28_decode"
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"timestamp": "2026-02-13T15:15:50.581622",
|
| 550 |
+
"timestamp_ms": 1770963350581.63,
|
| 551 |
+
"phase": "inference",
|
| 552 |
+
"sub_phase": "req_28_idle_between"
|
| 553 |
+
},
|
| 554 |
+
{
|
| 555 |
+
"timestamp": "2026-02-13T15:15:55.581916",
|
| 556 |
+
"timestamp_ms": 1770963355581.9514,
|
| 557 |
+
"phase": "inference",
|
| 558 |
+
"sub_phase": "req_29_prefill"
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"timestamp": "2026-02-13T15:15:55.600418",
|
| 562 |
+
"timestamp_ms": 1770963355600.4287,
|
| 563 |
+
"phase": "inference",
|
| 564 |
+
"sub_phase": "req_29_decode"
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"timestamp": "2026-02-13T15:15:57.363801",
|
| 568 |
+
"timestamp_ms": 1770963357363.8093,
|
| 569 |
+
"phase": "inference",
|
| 570 |
+
"sub_phase": "session_finished"
|
| 571 |
+
},
|
| 572 |
+
{
|
| 573 |
+
"timestamp": "2026-02-13T15:15:57.363902",
|
| 574 |
+
"timestamp_ms": 1770963357363.9043,
|
| 575 |
+
"phase": "idle_post",
|
| 576 |
+
"sub_phase": ""
|
| 577 |
+
}
|
| 578 |
+
]
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv
ADDED
|
@@ -0,0 +1,71 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
request_idx,input_tokens,output_tokens,prefill_time_ms,decode_time_ms,prompt_preview,output_preview
|
| 2 |
+
0,8,128,320.8603858947754,1936.0580444335938,Generate a list of random words.," Here is one list of 10 random words:
|
| 3 |
+
|
| 4 |
+
1. Space
|
| 5 |
+
2. Helicopter
|
| 6 |
+
3. Pineapple
|
| 7 |
+
4. G"
|
| 8 |
+
1,8,128,15.880346298217773,1757.9174041748047,Describe the structure of an atom.,"
|
| 9 |
+
An atom is a tiny particle that makes up everything around us, including the c"
|
| 10 |
+
2,8,128,16.511917114257812,1758.1751346588135,Describe the function of a computer motherboard,"
|
| 11 |
+
A computer motherboard is the primary circuit board of a computer, connecting a"
|
| 12 |
+
3,18,128,17.695188522338867,1761.1734867095947,"Explain why the following fraction is equivalent to 1/4
|
| 13 |
+
4/16"," is equivalent to 1/4
|
| 14 |
+
To see why, let’s try to reduce the fraction 4/16 to its l"
|
| 15 |
+
4,12,128,16.199588775634766,1758.7261199951172,Explain the process of cellular respiration in plants., Cellular respiration is the process by which cells break down glucose to produc
|
| 16 |
+
5,29,128,16.67642593383789,1760.8389854431152,"Given a set of numbers, find the maximum value.
|
| 17 |
+
Set: {10, 3, 25, 6, 16}","
|
| 18 |
+
First, we sort the set in ascending order: {3, 6, 10, 16, 25}
|
| 19 |
+
Then, we find th"
|
| 20 |
+
6,9,128,17.26984977722168,1759.8199844360352,Generate a poem with 10 lines.," The poem should be about the beauty of the stars in the night sky.
|
| 21 |
+
|
| 22 |
+
The stars s"
|
| 23 |
+
7,7,128,16.423940658569336,1759.322166442871,What does DNA stand for?," What does it represent?
|
| 24 |
+
A. DNA stands for Deoxyribonucleic Acid. It represents "
|
| 25 |
+
8,12,128,17.035722732543945,1760.2012157440186,What is the product of 6 and 2?,"
|
| 26 |
+
Step 1: To find the product of 6 and 2, we need to multiply these two numbers"
|
| 27 |
+
9,7,128,15.502452850341797,1760.1375579833984,How did Julius Caesar die?, Julius Caesar was a Roman general and statesman who played a critical role in t
|
| 28 |
+
10,30,128,17.156600952148438,1761.343002319336,"Analyze the given text for its tone.
|
| 29 |
+
The world has been greatly impacted by the "," We have had to adapt to a new normal, which includes social distancing, wearing"
|
| 30 |
+
11,22,128,17.437219619750977,1758.486032485962,"Arrange the given numbers in ascending order.
|
| 31 |
+
2, 4, 0, 8, 3",", 1, 6, 7, 5, 9
|
| 32 |
+
The correct answer is: 0, 1, 2, 3, 4, 5, 6, 7, 8, 9
|
| 33 |
+
The correct "
|
| 34 |
+
12,8,128,15.615463256835938,1755.5911540985107,Give three tips for staying healthy., I want to know what foods you recommend and what exercise routines you recommen
|
| 35 |
+
13,13,128,16.29328727722168,1755.5346488952637,Compare and contrast the Cuban Missile Crisis and the Vietnam War.," Essay Example
|
| 36 |
+
Compare and contrast the Cuban Missile Crisis and the Vietnam War"
|
| 37 |
+
14,15,128,15.985488891601562,1757.5948238372803,Generate a list of ten items a person might need for a camping trip,. This list should include items that are essential for a safe and comfortable c
|
| 38 |
+
15,8,128,16.08872413635254,1755.9077739715576,What are the three primary colors?, Primary colors are the basic colors that cannot be created by mixing other colo
|
| 39 |
+
16,23,128,16.10112190246582,1757.4870586395264,Variable x is defined as “4x + 2y = 10”. Find the value of x.,"
|
| 40 |
+
Step 1: Identify the variables and the equation.
|
| 41 |
+
The variable x is defined as"
|
| 42 |
+
17,8,128,16.155004501342773,1757.869005203247,Give two examples of a liquid.," Explain why they are liquids.
|
| 43 |
+
## Step 1: Identify the first example of a liquid"
|
| 44 |
+
18,72,128,38.91324996948242,1765.0864124298096,"Summarize the given passage.
|
| 45 |
+
A recent study showed that global climate change is", The study highlighted the need for immediate action to reduce greenhouse gas em
|
| 46 |
+
19,10,128,15.765905380249023,1757.2834491729736,Who is the world's most famous painter?," This is a subjective question, as opinions on art are often personal and varied"
|
| 47 |
+
20,11,128,16.478300094604492,1757.4574947357178,Describe the life and reign of King Charles II.," Charles II, also known as the Merry Monarch, was the king of England, Scotland,"
|
| 48 |
+
21,14,128,15.909910202026367,1759.0091228485107,"Transcribe the recording into text.
|
| 49 |
+
Recording about the history of Waterloo"," Region in Ontario, Canada.
|
| 50 |
+
Transcription:
|
| 51 |
+
Here's a brief history of Waterloo Re"
|
| 52 |
+
22,9,128,15.495061874389648,1759.9248886108398,Provide one example for a cultural practice., Provide one example for a cultural product. Provide one example for a cultural
|
| 53 |
+
23,19,128,16.555309295654297,1761.4634037017822,"Rewrite the following sentence using active voice.
|
| 54 |
+
The news report was read by t"," (passive voice)
|
| 55 |
+
The captain read the news report. (active voice)
|
| 56 |
+
In the passive"
|
| 57 |
+
24,8,128,16.040563583374023,1759.0117454528809,What is the capital of France?," Paris is the capital of France.
|
| 58 |
+
What is the largest city in the United States? "
|
| 59 |
+
25,13,128,15.721321105957031,1757.5478553771973,"Identify the odd one out.
|
| 60 |
+
Twitter, Instagram, Telegram",", Snapchat, Facebook
|
| 61 |
+
The odd one out is Telegram.
|
| 62 |
+
Why? Because the other four ar"
|
| 63 |
+
26,9,128,16.607999801635742,1760.4961395263672,Design an app for a delivery company.," Create a tracking system and provide information about the delivery status.
|
| 64 |
+
Des"
|
| 65 |
+
27,10,128,15.795230865478516,1759.629726409912,Brainstorm possible solutions to reduce water pollution.," How can we reduce the amount of pollution in our water sources?
|
| 66 |
+
Here are some p"
|
| 67 |
+
28,25,128,16.43657684326172,1761.1958980560303,"Use the given data to calculate the median.
|
| 68 |
+
[2, 3, 7, 8, 10]", (a) Find the median of the data set. (b) Find the median of the data set if 1 i
|
| 69 |
+
29,20,128,17.00758934020996,1763.085126876831,"Classify the following into animals, plants, and minerals
|
| 70 |
+
Oak tree, copper ore, ",", granite, dog, pine tree, gold, iron ore, cat, rose bush, diamond, rabbit
|
| 71 |
+
Anima"
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_llama3.1-8b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3324,33.357168,303.837636,9.110901,14.496,14.834,7.521
|
| 3 |
+
idle_post,3000,29.989999,1955.323326,65.179907,126.326,421.607,7.464
|
| 4 |
+
idle_pre,3001,30.010115,1610.136846,53.653079,78.186,80.069,7.742
|
| 5 |
+
inference_decode,5302,53.018674,12435.561062,234.545182,422.176,423.328,8.458
|
| 6 |
+
inference_idle,14500,145.000785,21469.096394,148.062656,421.073,423.328,77.506
|
| 7 |
+
inference_prefill,87,0.870546,44.165622,50.71892,78.131,79.661,7.979
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 3,
|
| 7 |
+
"train_max_steps": null,
|
| 8 |
+
"batch_size": 4,
|
| 9 |
+
"gradient_accumulation_steps": 4,
|
| 10 |
+
"learning_rate": 0.0002,
|
| 11 |
+
"lora_r": 16,
|
| 12 |
+
"lora_alpha": 32,
|
| 13 |
+
"lora_dropout": 0.05,
|
| 14 |
+
"max_seq_length": 512,
|
| 15 |
+
"use_4bit": false,
|
| 16 |
+
"use_bf16": true,
|
| 17 |
+
"inference_total_requests": 30,
|
| 18 |
+
"inference_interval_sec": 5.0,
|
| 19 |
+
"inference_pattern": "fixed",
|
| 20 |
+
"inference_variable_min": 2.0,
|
| 21 |
+
"inference_variable_max": 10.0,
|
| 22 |
+
"inference_burst_count": 3,
|
| 23 |
+
"inference_burst_gap": 1.0,
|
| 24 |
+
"inference_burst_pause": 15.0,
|
| 25 |
+
"gen_max_new_tokens": 128,
|
| 26 |
+
"gen_temperature": 0.7,
|
| 27 |
+
"gen_do_sample": true,
|
| 28 |
+
"prompt_set": "alpaca",
|
| 29 |
+
"clock_lock_gpu": null,
|
| 30 |
+
"clock_lock_mem": null,
|
| 31 |
+
"ramp_enabled": false,
|
| 32 |
+
"ramp_start_w": null,
|
| 33 |
+
"ramp_end_w": null,
|
| 34 |
+
"ramp_steps": 5,
|
| 35 |
+
"ramp_step_duration": 3.0,
|
| 36 |
+
"device": "cuda:0",
|
| 37 |
+
"gpu_index": 0,
|
| 38 |
+
"skip_training": true,
|
| 39 |
+
"skip_inference": false,
|
| 40 |
+
"model": "mistral-7b",
|
| 41 |
+
"power_cap": null,
|
| 42 |
+
"gpu_info": {
|
| 43 |
+
"gpu_name": "NVIDIA RTX PRO 6000 Blackwell Workstation Edition",
|
| 44 |
+
"tdp_w": 600,
|
| 45 |
+
"memory_total_mb": 97887,
|
| 46 |
+
"driver_version": "580.82.07",
|
| 47 |
+
"cuda_version": "13.0"
|
| 48 |
+
},
|
| 49 |
+
"experiment_timestamp": "2026-02-13T15:07:05.752612"
|
| 50 |
+
}
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json
ADDED
|
@@ -0,0 +1,578 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-13T15:02:18.701404",
|
| 4 |
+
"timestamp_ms": 1770962538701.4155,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-13T15:02:48.701917",
|
| 10 |
+
"timestamp_ms": 1770962568701.9565,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "skipped"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-13T15:02:48.702056",
|
| 16 |
+
"timestamp_ms": 1770962568702.0647,
|
| 17 |
+
"phase": "idle_mid",
|
| 18 |
+
"sub_phase": ""
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-13T15:03:18.702386",
|
| 22 |
+
"timestamp_ms": 1770962598702.4265,
|
| 23 |
+
"phase": "inference",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-13T15:03:21.785388",
|
| 28 |
+
"timestamp_ms": 1770962601785.4058,
|
| 29 |
+
"phase": "inference",
|
| 30 |
+
"sub_phase": "session_start_pattern_fixed"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-13T15:03:21.785519",
|
| 34 |
+
"timestamp_ms": 1770962601785.5222,
|
| 35 |
+
"phase": "inference",
|
| 36 |
+
"sub_phase": "req_0_prefill"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-13T15:03:22.082446",
|
| 40 |
+
"timestamp_ms": 1770962602082.4575,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "req_0_decode"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-13T15:03:23.939037",
|
| 46 |
+
"timestamp_ms": 1770962603939.047,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "req_0_idle_between"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-13T15:03:28.939283",
|
| 52 |
+
"timestamp_ms": 1770962608939.3206,
|
| 53 |
+
"phase": "inference",
|
| 54 |
+
"sub_phase": "req_1_prefill"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-13T15:03:28.958472",
|
| 58 |
+
"timestamp_ms": 1770962608958.4841,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "req_1_decode"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-13T15:03:30.631345",
|
| 64 |
+
"timestamp_ms": 1770962610631.3533,
|
| 65 |
+
"phase": "inference",
|
| 66 |
+
"sub_phase": "req_1_idle_between"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-13T15:03:35.631590",
|
| 70 |
+
"timestamp_ms": 1770962615631.6235,
|
| 71 |
+
"phase": "inference",
|
| 72 |
+
"sub_phase": "req_2_prefill"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-13T15:03:35.648148",
|
| 76 |
+
"timestamp_ms": 1770962615648.159,
|
| 77 |
+
"phase": "inference",
|
| 78 |
+
"sub_phase": "req_2_decode"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-13T15:03:37.316512",
|
| 82 |
+
"timestamp_ms": 1770962617316.5205,
|
| 83 |
+
"phase": "inference",
|
| 84 |
+
"sub_phase": "req_2_idle_between"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-13T15:03:42.316741",
|
| 88 |
+
"timestamp_ms": 1770962622316.7761,
|
| 89 |
+
"phase": "inference",
|
| 90 |
+
"sub_phase": "req_3_prefill"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-13T15:03:42.333953",
|
| 94 |
+
"timestamp_ms": 1770962622333.9639,
|
| 95 |
+
"phase": "inference",
|
| 96 |
+
"sub_phase": "req_3_decode"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-13T15:03:44.000210",
|
| 100 |
+
"timestamp_ms": 1770962624000.218,
|
| 101 |
+
"phase": "inference",
|
| 102 |
+
"sub_phase": "req_3_idle_between"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-13T15:03:49.000464",
|
| 106 |
+
"timestamp_ms": 1770962629000.4985,
|
| 107 |
+
"phase": "inference",
|
| 108 |
+
"sub_phase": "req_4_prefill"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-13T15:03:49.018062",
|
| 112 |
+
"timestamp_ms": 1770962629018.0732,
|
| 113 |
+
"phase": "inference",
|
| 114 |
+
"sub_phase": "req_4_decode"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-13T15:03:50.684286",
|
| 118 |
+
"timestamp_ms": 1770962630684.2937,
|
| 119 |
+
"phase": "inference",
|
| 120 |
+
"sub_phase": "req_4_idle_between"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-13T15:03:55.684541",
|
| 124 |
+
"timestamp_ms": 1770962635684.576,
|
| 125 |
+
"phase": "inference",
|
| 126 |
+
"sub_phase": "req_5_prefill"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-13T15:03:55.702317",
|
| 130 |
+
"timestamp_ms": 1770962635702.3271,
|
| 131 |
+
"phase": "inference",
|
| 132 |
+
"sub_phase": "req_5_decode"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-13T15:03:57.371243",
|
| 136 |
+
"timestamp_ms": 1770962637371.2507,
|
| 137 |
+
"phase": "inference",
|
| 138 |
+
"sub_phase": "req_5_idle_between"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-13T15:04:02.371476",
|
| 142 |
+
"timestamp_ms": 1770962642371.5127,
|
| 143 |
+
"phase": "inference",
|
| 144 |
+
"sub_phase": "req_6_prefill"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-13T15:04:02.388306",
|
| 148 |
+
"timestamp_ms": 1770962642388.3162,
|
| 149 |
+
"phase": "inference",
|
| 150 |
+
"sub_phase": "req_6_decode"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-13T15:04:04.056673",
|
| 154 |
+
"timestamp_ms": 1770962644056.6812,
|
| 155 |
+
"phase": "inference",
|
| 156 |
+
"sub_phase": "req_6_idle_between"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-13T15:04:09.056990",
|
| 160 |
+
"timestamp_ms": 1770962649057.026,
|
| 161 |
+
"phase": "inference",
|
| 162 |
+
"sub_phase": "req_7_prefill"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-13T15:04:09.073491",
|
| 166 |
+
"timestamp_ms": 1770962649073.5027,
|
| 167 |
+
"phase": "inference",
|
| 168 |
+
"sub_phase": "req_7_decode"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-13T15:04:10.740891",
|
| 172 |
+
"timestamp_ms": 1770962650740.9001,
|
| 173 |
+
"phase": "inference",
|
| 174 |
+
"sub_phase": "req_7_idle_between"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-13T15:04:15.741167",
|
| 178 |
+
"timestamp_ms": 1770962655741.205,
|
| 179 |
+
"phase": "inference",
|
| 180 |
+
"sub_phase": "req_8_prefill"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-13T15:04:15.758526",
|
| 184 |
+
"timestamp_ms": 1770962655758.5378,
|
| 185 |
+
"phase": "inference",
|
| 186 |
+
"sub_phase": "req_8_decode"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-13T15:04:17.429929",
|
| 190 |
+
"timestamp_ms": 1770962657429.9373,
|
| 191 |
+
"phase": "inference",
|
| 192 |
+
"sub_phase": "req_8_idle_between"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-13T15:04:22.430154",
|
| 196 |
+
"timestamp_ms": 1770962662430.1887,
|
| 197 |
+
"phase": "inference",
|
| 198 |
+
"sub_phase": "req_9_prefill"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-13T15:04:22.446506",
|
| 202 |
+
"timestamp_ms": 1770962662446.5166,
|
| 203 |
+
"phase": "inference",
|
| 204 |
+
"sub_phase": "req_9_decode"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-13T15:04:24.115520",
|
| 208 |
+
"timestamp_ms": 1770962664115.5276,
|
| 209 |
+
"phase": "inference",
|
| 210 |
+
"sub_phase": "req_9_idle_between"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-13T15:04:29.115790",
|
| 214 |
+
"timestamp_ms": 1770962669115.824,
|
| 215 |
+
"phase": "inference",
|
| 216 |
+
"sub_phase": "req_10_prefill"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-13T15:04:29.132790",
|
| 220 |
+
"timestamp_ms": 1770962669132.801,
|
| 221 |
+
"phase": "inference",
|
| 222 |
+
"sub_phase": "req_10_decode"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-13T15:04:30.802120",
|
| 226 |
+
"timestamp_ms": 1770962670802.1282,
|
| 227 |
+
"phase": "inference",
|
| 228 |
+
"sub_phase": "req_10_idle_between"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-13T15:04:35.802351",
|
| 232 |
+
"timestamp_ms": 1770962675802.3855,
|
| 233 |
+
"phase": "inference",
|
| 234 |
+
"sub_phase": "req_11_prefill"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-13T15:04:35.819136",
|
| 238 |
+
"timestamp_ms": 1770962675819.1462,
|
| 239 |
+
"phase": "inference",
|
| 240 |
+
"sub_phase": "req_11_decode"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-13T15:04:37.488267",
|
| 244 |
+
"timestamp_ms": 1770962677488.2751,
|
| 245 |
+
"phase": "inference",
|
| 246 |
+
"sub_phase": "req_11_idle_between"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-13T15:04:42.488497",
|
| 250 |
+
"timestamp_ms": 1770962682488.5337,
|
| 251 |
+
"phase": "inference",
|
| 252 |
+
"sub_phase": "req_12_prefill"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-13T15:04:42.506300",
|
| 256 |
+
"timestamp_ms": 1770962682506.3125,
|
| 257 |
+
"phase": "inference",
|
| 258 |
+
"sub_phase": "req_12_decode"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-13T15:04:44.175925",
|
| 262 |
+
"timestamp_ms": 1770962684175.9326,
|
| 263 |
+
"phase": "inference",
|
| 264 |
+
"sub_phase": "req_12_idle_between"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-13T15:04:49.176188",
|
| 268 |
+
"timestamp_ms": 1770962689176.222,
|
| 269 |
+
"phase": "inference",
|
| 270 |
+
"sub_phase": "req_13_prefill"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-13T15:04:49.193845",
|
| 274 |
+
"timestamp_ms": 1770962689193.8564,
|
| 275 |
+
"phase": "inference",
|
| 276 |
+
"sub_phase": "req_13_decode"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-13T15:04:50.863607",
|
| 280 |
+
"timestamp_ms": 1770962690863.615,
|
| 281 |
+
"phase": "inference",
|
| 282 |
+
"sub_phase": "req_13_idle_between"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-13T15:04:55.863899",
|
| 286 |
+
"timestamp_ms": 1770962695863.9333,
|
| 287 |
+
"phase": "inference",
|
| 288 |
+
"sub_phase": "req_14_prefill"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-13T15:04:55.882198",
|
| 292 |
+
"timestamp_ms": 1770962695882.209,
|
| 293 |
+
"phase": "inference",
|
| 294 |
+
"sub_phase": "req_14_decode"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-13T15:04:57.551041",
|
| 298 |
+
"timestamp_ms": 1770962697551.0496,
|
| 299 |
+
"phase": "inference",
|
| 300 |
+
"sub_phase": "req_14_idle_between"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-13T15:05:02.551270",
|
| 304 |
+
"timestamp_ms": 1770962702551.3066,
|
| 305 |
+
"phase": "inference",
|
| 306 |
+
"sub_phase": "req_15_prefill"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-13T15:05:02.569273",
|
| 310 |
+
"timestamp_ms": 1770962702569.2842,
|
| 311 |
+
"phase": "inference",
|
| 312 |
+
"sub_phase": "req_15_decode"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-13T15:05:04.237536",
|
| 316 |
+
"timestamp_ms": 1770962704237.5437,
|
| 317 |
+
"phase": "inference",
|
| 318 |
+
"sub_phase": "req_15_idle_between"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-13T15:05:09.237808",
|
| 322 |
+
"timestamp_ms": 1770962709237.8855,
|
| 323 |
+
"phase": "inference",
|
| 324 |
+
"sub_phase": "req_16_prefill"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-13T15:05:09.256302",
|
| 328 |
+
"timestamp_ms": 1770962709256.3132,
|
| 329 |
+
"phase": "inference",
|
| 330 |
+
"sub_phase": "req_16_decode"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-13T15:05:10.927394",
|
| 334 |
+
"timestamp_ms": 1770962710927.4019,
|
| 335 |
+
"phase": "inference",
|
| 336 |
+
"sub_phase": "req_16_idle_between"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-13T15:05:15.927650",
|
| 340 |
+
"timestamp_ms": 1770962715927.6853,
|
| 341 |
+
"phase": "inference",
|
| 342 |
+
"sub_phase": "req_17_prefill"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-13T15:05:15.946069",
|
| 346 |
+
"timestamp_ms": 1770962715946.08,
|
| 347 |
+
"phase": "inference",
|
| 348 |
+
"sub_phase": "req_17_decode"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-13T15:05:17.617824",
|
| 352 |
+
"timestamp_ms": 1770962717617.8342,
|
| 353 |
+
"phase": "inference",
|
| 354 |
+
"sub_phase": "req_17_idle_between"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-13T15:05:22.618077",
|
| 358 |
+
"timestamp_ms": 1770962722618.1138,
|
| 359 |
+
"phase": "inference",
|
| 360 |
+
"sub_phase": "req_18_prefill"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-13T15:05:22.637888",
|
| 364 |
+
"timestamp_ms": 1770962722637.9006,
|
| 365 |
+
"phase": "inference",
|
| 366 |
+
"sub_phase": "req_18_decode"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-13T15:05:24.308184",
|
| 370 |
+
"timestamp_ms": 1770962724308.1926,
|
| 371 |
+
"phase": "inference",
|
| 372 |
+
"sub_phase": "req_18_idle_between"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-13T15:05:29.308432",
|
| 376 |
+
"timestamp_ms": 1770962729308.4646,
|
| 377 |
+
"phase": "inference",
|
| 378 |
+
"sub_phase": "req_19_prefill"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-13T15:05:29.325781",
|
| 382 |
+
"timestamp_ms": 1770962729325.7917,
|
| 383 |
+
"phase": "inference",
|
| 384 |
+
"sub_phase": "req_19_decode"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-13T15:05:30.996919",
|
| 388 |
+
"timestamp_ms": 1770962730996.927,
|
| 389 |
+
"phase": "inference",
|
| 390 |
+
"sub_phase": "req_19_idle_between"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-13T15:05:35.997153",
|
| 394 |
+
"timestamp_ms": 1770962735997.1848,
|
| 395 |
+
"phase": "inference",
|
| 396 |
+
"sub_phase": "req_20_prefill"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-13T15:05:36.014776",
|
| 400 |
+
"timestamp_ms": 1770962736014.7856,
|
| 401 |
+
"phase": "inference",
|
| 402 |
+
"sub_phase": "req_20_decode"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-13T15:05:37.682295",
|
| 406 |
+
"timestamp_ms": 1770962737682.3027,
|
| 407 |
+
"phase": "inference",
|
| 408 |
+
"sub_phase": "req_20_idle_between"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-13T15:05:42.682496",
|
| 412 |
+
"timestamp_ms": 1770962742682.5288,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "req_21_prefill"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-13T15:05:42.700536",
|
| 418 |
+
"timestamp_ms": 1770962742700.5454,
|
| 419 |
+
"phase": "inference",
|
| 420 |
+
"sub_phase": "req_21_decode"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-13T15:05:44.370094",
|
| 424 |
+
"timestamp_ms": 1770962744370.102,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "req_21_idle_between"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-13T15:05:49.370341",
|
| 430 |
+
"timestamp_ms": 1770962749370.374,
|
| 431 |
+
"phase": "inference",
|
| 432 |
+
"sub_phase": "req_22_prefill"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-13T15:05:49.387027",
|
| 436 |
+
"timestamp_ms": 1770962749387.0361,
|
| 437 |
+
"phase": "inference",
|
| 438 |
+
"sub_phase": "req_22_decode"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-13T15:05:51.080188",
|
| 442 |
+
"timestamp_ms": 1770962751080.196,
|
| 443 |
+
"phase": "inference",
|
| 444 |
+
"sub_phase": "req_22_idle_between"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-13T15:05:56.080414",
|
| 448 |
+
"timestamp_ms": 1770962756080.448,
|
| 449 |
+
"phase": "inference",
|
| 450 |
+
"sub_phase": "req_23_prefill"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-13T15:05:56.098375",
|
| 454 |
+
"timestamp_ms": 1770962756098.3855,
|
| 455 |
+
"phase": "inference",
|
| 456 |
+
"sub_phase": "req_23_decode"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-13T15:05:57.766889",
|
| 460 |
+
"timestamp_ms": 1770962757766.897,
|
| 461 |
+
"phase": "inference",
|
| 462 |
+
"sub_phase": "req_23_idle_between"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-13T15:06:02.767109",
|
| 466 |
+
"timestamp_ms": 1770962762767.1458,
|
| 467 |
+
"phase": "inference",
|
| 468 |
+
"sub_phase": "req_24_prefill"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-13T15:06:02.784020",
|
| 472 |
+
"timestamp_ms": 1770962762784.0308,
|
| 473 |
+
"phase": "inference",
|
| 474 |
+
"sub_phase": "req_24_decode"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-13T15:06:04.451218",
|
| 478 |
+
"timestamp_ms": 1770962764451.2258,
|
| 479 |
+
"phase": "inference",
|
| 480 |
+
"sub_phase": "req_24_idle_between"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-13T15:06:09.451479",
|
| 484 |
+
"timestamp_ms": 1770962769451.5156,
|
| 485 |
+
"phase": "inference",
|
| 486 |
+
"sub_phase": "req_25_prefill"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-13T15:06:09.468864",
|
| 490 |
+
"timestamp_ms": 1770962769468.8743,
|
| 491 |
+
"phase": "inference",
|
| 492 |
+
"sub_phase": "req_25_decode"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-13T15:06:10.435038",
|
| 496 |
+
"timestamp_ms": 1770962770435.045,
|
| 497 |
+
"phase": "inference",
|
| 498 |
+
"sub_phase": "req_25_idle_between"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-13T15:06:15.435300",
|
| 502 |
+
"timestamp_ms": 1770962775435.3352,
|
| 503 |
+
"phase": "inference",
|
| 504 |
+
"sub_phase": "req_26_prefill"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-13T15:06:15.452479",
|
| 508 |
+
"timestamp_ms": 1770962775452.4895,
|
| 509 |
+
"phase": "inference",
|
| 510 |
+
"sub_phase": "req_26_decode"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-13T15:06:16.784189",
|
| 514 |
+
"timestamp_ms": 1770962776784.1973,
|
| 515 |
+
"phase": "inference",
|
| 516 |
+
"sub_phase": "req_26_idle_between"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-13T15:06:21.784475",
|
| 520 |
+
"timestamp_ms": 1770962781784.5151,
|
| 521 |
+
"phase": "inference",
|
| 522 |
+
"sub_phase": "req_27_prefill"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-13T15:06:21.802665",
|
| 526 |
+
"timestamp_ms": 1770962781802.6807,
|
| 527 |
+
"phase": "inference",
|
| 528 |
+
"sub_phase": "req_27_decode"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-13T15:06:22.314857",
|
| 532 |
+
"timestamp_ms": 1770962782314.8652,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "req_27_idle_between"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-13T15:06:27.315154",
|
| 538 |
+
"timestamp_ms": 1770962787315.1946,
|
| 539 |
+
"phase": "inference",
|
| 540 |
+
"sub_phase": "req_28_prefill"
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"timestamp": "2026-02-13T15:06:27.372821",
|
| 544 |
+
"timestamp_ms": 1770962787372.8333,
|
| 545 |
+
"phase": "inference",
|
| 546 |
+
"sub_phase": "req_28_decode"
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"timestamp": "2026-02-13T15:06:29.045190",
|
| 550 |
+
"timestamp_ms": 1770962789045.1992,
|
| 551 |
+
"phase": "inference",
|
| 552 |
+
"sub_phase": "req_28_idle_between"
|
| 553 |
+
},
|
| 554 |
+
{
|
| 555 |
+
"timestamp": "2026-02-13T15:06:34.045431",
|
| 556 |
+
"timestamp_ms": 1770962794045.466,
|
| 557 |
+
"phase": "inference",
|
| 558 |
+
"sub_phase": "req_29_prefill"
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"timestamp": "2026-02-13T15:06:34.063477",
|
| 562 |
+
"timestamp_ms": 1770962794063.4878,
|
| 563 |
+
"phase": "inference",
|
| 564 |
+
"sub_phase": "req_29_decode"
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"timestamp": "2026-02-13T15:06:35.735967",
|
| 568 |
+
"timestamp_ms": 1770962795735.9753,
|
| 569 |
+
"phase": "inference",
|
| 570 |
+
"sub_phase": "session_finished"
|
| 571 |
+
},
|
| 572 |
+
{
|
| 573 |
+
"timestamp": "2026-02-13T15:06:35.736066",
|
| 574 |
+
"timestamp_ms": 1770962795736.0676,
|
| 575 |
+
"phase": "idle_post",
|
| 576 |
+
"sub_phase": ""
|
| 577 |
+
}
|
| 578 |
+
]
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv
ADDED
|
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
request_idx,input_tokens,output_tokens,prefill_time_ms,decode_time_ms,prompt_preview,output_preview
|
| 2 |
+
0,35,128,279.5259952545166,1855.9198379516602,"Analyze the given text for its tone.
|
| 3 |
+
The world has been greatly impacted by the ","As we continue to adapt to this “new normal,” it is important to remember the le"
|
| 4 |
+
1,25,128,17.602205276489258,1672.5170612335205,Variable x is defined as “4x + 2y = 10”. Find the value of x.,"
|
| 5 |
+
|
| 6 |
+
First, we need to solve for x. To do this, we can subtract 2y from both sides "
|
| 7 |
+
2,14,128,15.091180801391602,1668.0312156677246,Explain the use of word embeddings in Natural Language Processing,"(NLP)
|
| 8 |
+
|
| 9 |
+
Word embeddings in Natural Language Processing (NLP) are a type of word r"
|
| 10 |
+
3,11,128,15.8843994140625,1665.9305095672607,Who is the world's most famous painter?,Leonardo da Vinci is considered the world's most famous painter. He is best know
|
| 11 |
+
4,13,128,16.185760498046875,1665.8780574798584,Describe a time when you had to make a difficult decision.,"
|
| 12 |
+
|
| 13 |
+
In the spring of 2019, I found myself in a challenging position, one that requ"
|
| 14 |
+
5,15,128,16.316652297973633,1668.6017513275146,"Generate three verbs that mean the same as ""to apologize""","
|
| 15 |
+
|
| 16 |
+
1. To express regret: This verb means to openly communicate that you feel sorr"
|
| 17 |
+
6,10,128,15.38538932800293,1668.036699295044,Discuss the causes of the Great Depression,".
|
| 18 |
+
|
| 19 |
+
The Great Depression was a severe economic downturn that occurred in the 1930"
|
| 20 |
+
7,13,128,15.118598937988281,1667.0458316802979,Explain the process of cellular respiration in plants.,"
|
| 21 |
+
|
| 22 |
+
Cellular respiration is a process by which cells convert biochemical energy fr"
|
| 23 |
+
8,12,128,15.792608261108398,1671.0522174835205,Explain the concept of cogging torque.,"
|
| 24 |
+
|
| 25 |
+
Cogging torque, also known as magneto-motive force (MMF) torque or detent torq"
|
| 26 |
+
9,10,128,14.977216720581055,1668.687105178833,Render a 3D model of a house,"and its surroundings, including vegetation, furniture, and people.
|
| 27 |
+
|
| 28 |
+
To create a "
|
| 29 |
+
10,16,128,15.583038330078125,1668.994426727295,"Identify the odd one out.
|
| 30 |
+
Twitter, Instagram, Telegram",", and Blogger. That’s a tough one. Let’s break it down: Twitter and Instagram ar"
|
| 31 |
+
11,8,128,15.421390533447266,1668.8275337219238,Generate a list of random words.,"
|
| 32 |
+
|
| 33 |
+
A new list of random words can be generated by using the `random_word_list` fu"
|
| 34 |
+
12,19,128,16.290903091430664,1669.273853302002,"Transcribe the recording into text.
|
| 35 |
+
Recording about the history of Waterloo",", Ontario
|
| 36 |
+
|
| 37 |
+
[Narrator]
|
| 38 |
+
In the heart of Southwestern Ontario lies the city of Wate"
|
| 39 |
+
13,12,128,16.2355899810791,1669.4316864013672,Explain the concept of artificial intelligence in simple terms.,"
|
| 40 |
+
|
| 41 |
+
Artificial Intelligence, often abbreviated as AI, is a field of computer scien"
|
| 42 |
+
14,18,128,16.89457893371582,1668.5261726379395,What is the force on a 1 kg mass due to the gravitational force?,"
|
| 43 |
+
|
| 44 |
+
The gravitational force on an object is given by the formula F = G * (m1 * m2)"
|
| 45 |
+
15,11,128,16.601085662841797,1667.9136753082275,Brainstorm possible solutions to reduce water pollution.,"
|
| 46 |
+
|
| 47 |
+
1. Implementing stricter regulations on industrial and agricultural waste disp"
|
| 48 |
+
16,21,128,17.008543014526367,1670.764446258545,"Rewrite the following sentence using active voice.
|
| 49 |
+
The news report was read by t","
|
| 50 |
+
|
| 51 |
+
The captain read the news report.
|
| 52 |
+
|
| 53 |
+
Rewrite the following sentence using passiv"
|
| 54 |
+
17,33,128,16.85643196105957,1671.424388885498,"Given a set of numbers, find the maximum value.
|
| 55 |
+
Set: {10, 3, 25, 6, 16}","
|
| 56 |
+
|
| 57 |
+
The maximum value is 25.
|
| 58 |
+
|
| 59 |
+
Given a set of integers, return the maximum value.
|
| 60 |
+
|
| 61 |
+
"
|
| 62 |
+
18,8,128,18.440961837768555,1669.9459552764893,How can we reduce air pollution?,"There are many ways to reduce air pollution, and it is important to take action "
|
| 63 |
+
19,22,128,16.04914665222168,1670.8288192749023,"Convert from celsius to fahrenheit.
|
| 64 |
+
Temperature in Celsius: 15","°C
|
| 65 |
+
Temperature in Fahrenheit: 59°F
|
| 66 |
+
|
| 67 |
+
Convert from celsius to kelvin.
|
| 68 |
+
Temperature "
|
| 69 |
+
20,11,128,16.39103889465332,1667.2108173370361,Design a logo for a website about recycling.,"The website focuses on the importance of recycling, the benefits of recycling, a"
|
| 70 |
+
21,18,128,16.678571701049805,1669.2430973052979,"Write a short paragraph about the given topic.
|
| 71 |
+
The importance of using renewable","cannot be overstated. Renewable energy sources, such as wind, solar, and hydro p"
|
| 72 |
+
22,9,128,15.42973518371582,1669.2485809326172,Generate a poem that expresses joy.,"
|
| 73 |
+
|
| 74 |
+
In the heart of the morning, where the sun does play,
|
| 75 |
+
A symphony of colors, in"
|
| 76 |
+
23,10,128,16.62421226501465,1668.191909790039,Outline the consequences of deforestation.,"
|
| 77 |
+
|
| 78 |
+
Deforestation has a significant impact on various aspects of the environment a"
|
| 79 |
+
24,9,128,15.437841415405273,1666.858196258545,Describe the function of a computer motherboard,".
|
| 80 |
+
|
| 81 |
+
A computer motherboard, often referred to as the mainboard, is the central ci"
|
| 82 |
+
25,16,74,15.859127044677734,965.8749103546143,"Find the area of a circle given its radius.
|
| 83 |
+
Radius = 4","cm
|
| 84 |
+
Area of a Circle = π × r^{2}
|
| 85 |
+
Area of the circle = π × (4)^{2}
|
| 86 |
+
= 16π square cm"
|
| 87 |
+
26,25,102,15.606403350830078,1331.3851356506348,"Evaluate this sentence for spelling and grammar mistakes
|
| 88 |
+
He finnished his meal a",".
|
| 89 |
+
The sentence has a spelling mistake in the word ""finnished"" and should be ""fin"
|
| 90 |
+
27,12,39,16.58177375793457,511.86060905456543,What is the product of 6 and 2?,"
|
| 91 |
+
|
| 92 |
+
The product of 6 and 2 is 12. To find the product, you can multiply the number"
|
| 93 |
+
28,12,128,15.995264053344727,1672.0077991485596,Develop a plan to reduce electricity usage in a home.,"
|
| 94 |
+
|
| 95 |
+
Title: Home Energy Efficiency Plan
|
| 96 |
+
|
| 97 |
+
1. Energy Audit:
|
| 98 |
+
- Hire a professional "
|
| 99 |
+
29,50,128,16.27182960510254,1672.1444129943848,"Extract the facts from the paragraph.
|
| 100 |
+
Online education continues to become more ","In addition, online education allows for a greater access to a diversity of educ"
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_mistral-7b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3309,33.089904,301.422803,9.1092,14.691,14.742,7.661
|
| 3 |
+
idle_post,3000,29.990008,1948.79249,64.962303,141.69,417.962,7.477
|
| 4 |
+
idle_pre,3000,30.00016,1586.395752,52.879611,78.576,80.624,7.932
|
| 5 |
+
inference_decode,4805,48.049664,10514.571696,218.827262,418.51,419.141,14.772
|
| 6 |
+
inference_idle,14501,145.011508,21375.647398,147.406668,417.747,419.141,77.349
|
| 7 |
+
inference_prefill,89,0.888826,50.032122,56.318944,78.317,78.467,10.193
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 3,
|
| 7 |
+
"train_max_steps": null,
|
| 8 |
+
"batch_size": 4,
|
| 9 |
+
"gradient_accumulation_steps": 4,
|
| 10 |
+
"learning_rate": 0.0002,
|
| 11 |
+
"lora_r": 16,
|
| 12 |
+
"lora_alpha": 32,
|
| 13 |
+
"lora_dropout": 0.05,
|
| 14 |
+
"max_seq_length": 512,
|
| 15 |
+
"use_4bit": false,
|
| 16 |
+
"use_bf16": true,
|
| 17 |
+
"inference_total_requests": 30,
|
| 18 |
+
"inference_interval_sec": 5.0,
|
| 19 |
+
"inference_pattern": "fixed",
|
| 20 |
+
"inference_variable_min": 2.0,
|
| 21 |
+
"inference_variable_max": 10.0,
|
| 22 |
+
"inference_burst_count": 3,
|
| 23 |
+
"inference_burst_gap": 1.0,
|
| 24 |
+
"inference_burst_pause": 15.0,
|
| 25 |
+
"gen_max_new_tokens": 128,
|
| 26 |
+
"gen_temperature": 0.7,
|
| 27 |
+
"gen_do_sample": true,
|
| 28 |
+
"prompt_set": "alpaca",
|
| 29 |
+
"clock_lock_gpu": null,
|
| 30 |
+
"clock_lock_mem": null,
|
| 31 |
+
"ramp_enabled": false,
|
| 32 |
+
"ramp_start_w": null,
|
| 33 |
+
"ramp_end_w": null,
|
| 34 |
+
"ramp_steps": 5,
|
| 35 |
+
"ramp_step_duration": 3.0,
|
| 36 |
+
"device": "cuda:0",
|
| 37 |
+
"gpu_index": 0,
|
| 38 |
+
"skip_training": true,
|
| 39 |
+
"skip_inference": false,
|
| 40 |
+
"model": "qwen3-4b",
|
| 41 |
+
"power_cap": null,
|
| 42 |
+
"gpu_info": {
|
| 43 |
+
"gpu_name": "NVIDIA RTX PRO 6000 Blackwell Workstation Edition",
|
| 44 |
+
"tdp_w": 600,
|
| 45 |
+
"memory_total_mb": 97887,
|
| 46 |
+
"driver_version": "580.82.07",
|
| 47 |
+
"cuda_version": "13.0"
|
| 48 |
+
},
|
| 49 |
+
"experiment_timestamp": "2026-02-13T15:00:07.417347"
|
| 50 |
+
}
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/phase_transitions.json
ADDED
|
@@ -0,0 +1,578 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-13T14:55:15.032368",
|
| 4 |
+
"timestamp_ms": 1770962115032.3792,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-13T14:55:45.032807",
|
| 10 |
+
"timestamp_ms": 1770962145032.8765,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "skipped"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-13T14:55:45.032990",
|
| 16 |
+
"timestamp_ms": 1770962145032.9985,
|
| 17 |
+
"phase": "idle_mid",
|
| 18 |
+
"sub_phase": ""
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-13T14:56:15.033249",
|
| 22 |
+
"timestamp_ms": 1770962175033.2903,
|
| 23 |
+
"phase": "inference",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-13T14:56:18.240965",
|
| 28 |
+
"timestamp_ms": 1770962178240.9778,
|
| 29 |
+
"phase": "inference",
|
| 30 |
+
"sub_phase": "session_start_pattern_fixed"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-13T14:56:18.241176",
|
| 34 |
+
"timestamp_ms": 1770962178241.1797,
|
| 35 |
+
"phase": "inference",
|
| 36 |
+
"sub_phase": "req_0_prefill"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-13T14:56:18.567272",
|
| 40 |
+
"timestamp_ms": 1770962178567.283,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "req_0_decode"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-13T14:56:20.532641",
|
| 46 |
+
"timestamp_ms": 1770962180532.651,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "req_0_idle_between"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-13T14:56:25.532937",
|
| 52 |
+
"timestamp_ms": 1770962185532.9724,
|
| 53 |
+
"phase": "inference",
|
| 54 |
+
"sub_phase": "req_1_prefill"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-13T14:56:25.553415",
|
| 58 |
+
"timestamp_ms": 1770962185553.4253,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "req_1_decode"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-13T14:56:27.318684",
|
| 64 |
+
"timestamp_ms": 1770962187318.6926,
|
| 65 |
+
"phase": "inference",
|
| 66 |
+
"sub_phase": "req_1_idle_between"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-13T14:56:32.318954",
|
| 70 |
+
"timestamp_ms": 1770962192318.9885,
|
| 71 |
+
"phase": "inference",
|
| 72 |
+
"sub_phase": "req_2_prefill"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-13T14:56:32.339031",
|
| 76 |
+
"timestamp_ms": 1770962192339.0403,
|
| 77 |
+
"phase": "inference",
|
| 78 |
+
"sub_phase": "req_2_decode"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-13T14:56:34.119111",
|
| 82 |
+
"timestamp_ms": 1770962194119.1187,
|
| 83 |
+
"phase": "inference",
|
| 84 |
+
"sub_phase": "req_2_idle_between"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-13T14:56:39.119345",
|
| 88 |
+
"timestamp_ms": 1770962199119.3801,
|
| 89 |
+
"phase": "inference",
|
| 90 |
+
"sub_phase": "req_3_prefill"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-13T14:56:39.138644",
|
| 94 |
+
"timestamp_ms": 1770962199138.6548,
|
| 95 |
+
"phase": "inference",
|
| 96 |
+
"sub_phase": "req_3_decode"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-13T14:56:40.906222",
|
| 100 |
+
"timestamp_ms": 1770962200906.23,
|
| 101 |
+
"phase": "inference",
|
| 102 |
+
"sub_phase": "req_3_idle_between"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-13T14:56:45.906472",
|
| 106 |
+
"timestamp_ms": 1770962205906.507,
|
| 107 |
+
"phase": "inference",
|
| 108 |
+
"sub_phase": "req_4_prefill"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-13T14:56:45.926409",
|
| 112 |
+
"timestamp_ms": 1770962205926.4202,
|
| 113 |
+
"phase": "inference",
|
| 114 |
+
"sub_phase": "req_4_decode"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-13T14:56:47.725945",
|
| 118 |
+
"timestamp_ms": 1770962207725.9531,
|
| 119 |
+
"phase": "inference",
|
| 120 |
+
"sub_phase": "req_4_idle_between"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-13T14:56:52.726197",
|
| 124 |
+
"timestamp_ms": 1770962212726.231,
|
| 125 |
+
"phase": "inference",
|
| 126 |
+
"sub_phase": "req_5_prefill"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-13T14:56:52.746121",
|
| 130 |
+
"timestamp_ms": 1770962212746.1318,
|
| 131 |
+
"phase": "inference",
|
| 132 |
+
"sub_phase": "req_5_decode"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-13T14:56:54.540175",
|
| 136 |
+
"timestamp_ms": 1770962214540.1829,
|
| 137 |
+
"phase": "inference",
|
| 138 |
+
"sub_phase": "req_5_idle_between"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-13T14:56:59.540418",
|
| 142 |
+
"timestamp_ms": 1770962219540.4502,
|
| 143 |
+
"phase": "inference",
|
| 144 |
+
"sub_phase": "req_6_prefill"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-13T14:56:59.559342",
|
| 148 |
+
"timestamp_ms": 1770962219559.351,
|
| 149 |
+
"phase": "inference",
|
| 150 |
+
"sub_phase": "req_6_decode"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-13T14:57:01.373282",
|
| 154 |
+
"timestamp_ms": 1770962221373.2905,
|
| 155 |
+
"phase": "inference",
|
| 156 |
+
"sub_phase": "req_6_idle_between"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-13T14:57:06.373535",
|
| 160 |
+
"timestamp_ms": 1770962226373.57,
|
| 161 |
+
"phase": "inference",
|
| 162 |
+
"sub_phase": "req_7_prefill"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-13T14:57:06.394087",
|
| 166 |
+
"timestamp_ms": 1770962226394.0986,
|
| 167 |
+
"phase": "inference",
|
| 168 |
+
"sub_phase": "req_7_decode"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-13T14:57:08.159595",
|
| 172 |
+
"timestamp_ms": 1770962228159.6028,
|
| 173 |
+
"phase": "inference",
|
| 174 |
+
"sub_phase": "req_7_idle_between"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-13T14:57:13.159897",
|
| 178 |
+
"timestamp_ms": 1770962233159.9333,
|
| 179 |
+
"phase": "inference",
|
| 180 |
+
"sub_phase": "req_8_prefill"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-13T14:57:13.178967",
|
| 184 |
+
"timestamp_ms": 1770962233178.9792,
|
| 185 |
+
"phase": "inference",
|
| 186 |
+
"sub_phase": "req_8_decode"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-13T14:57:14.963685",
|
| 190 |
+
"timestamp_ms": 1770962234963.693,
|
| 191 |
+
"phase": "inference",
|
| 192 |
+
"sub_phase": "req_8_idle_between"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-13T14:57:19.963980",
|
| 196 |
+
"timestamp_ms": 1770962239964.0137,
|
| 197 |
+
"phase": "inference",
|
| 198 |
+
"sub_phase": "req_9_prefill"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-13T14:57:19.983577",
|
| 202 |
+
"timestamp_ms": 1770962239983.5872,
|
| 203 |
+
"phase": "inference",
|
| 204 |
+
"sub_phase": "req_9_decode"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-13T14:57:21.767983",
|
| 208 |
+
"timestamp_ms": 1770962241767.9905,
|
| 209 |
+
"phase": "inference",
|
| 210 |
+
"sub_phase": "req_9_idle_between"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-13T14:57:26.768197",
|
| 214 |
+
"timestamp_ms": 1770962246768.232,
|
| 215 |
+
"phase": "inference",
|
| 216 |
+
"sub_phase": "req_10_prefill"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-13T14:57:26.787421",
|
| 220 |
+
"timestamp_ms": 1770962246787.432,
|
| 221 |
+
"phase": "inference",
|
| 222 |
+
"sub_phase": "req_10_decode"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-13T14:57:28.571056",
|
| 226 |
+
"timestamp_ms": 1770962248571.064,
|
| 227 |
+
"phase": "inference",
|
| 228 |
+
"sub_phase": "req_10_idle_between"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-13T14:57:33.571302",
|
| 232 |
+
"timestamp_ms": 1770962253571.337,
|
| 233 |
+
"phase": "inference",
|
| 234 |
+
"sub_phase": "req_11_prefill"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-13T14:57:33.589871",
|
| 238 |
+
"timestamp_ms": 1770962253589.8823,
|
| 239 |
+
"phase": "inference",
|
| 240 |
+
"sub_phase": "req_11_decode"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-13T14:57:35.373987",
|
| 244 |
+
"timestamp_ms": 1770962255373.9954,
|
| 245 |
+
"phase": "inference",
|
| 246 |
+
"sub_phase": "req_11_idle_between"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-13T14:57:40.374233",
|
| 250 |
+
"timestamp_ms": 1770962260374.2678,
|
| 251 |
+
"phase": "inference",
|
| 252 |
+
"sub_phase": "req_12_prefill"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-13T14:57:40.393158",
|
| 256 |
+
"timestamp_ms": 1770962260393.1694,
|
| 257 |
+
"phase": "inference",
|
| 258 |
+
"sub_phase": "req_12_decode"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-13T14:57:42.177021",
|
| 262 |
+
"timestamp_ms": 1770962262177.0293,
|
| 263 |
+
"phase": "inference",
|
| 264 |
+
"sub_phase": "req_12_idle_between"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-13T14:57:47.177269",
|
| 268 |
+
"timestamp_ms": 1770962267177.3042,
|
| 269 |
+
"phase": "inference",
|
| 270 |
+
"sub_phase": "req_13_prefill"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-13T14:57:47.196627",
|
| 274 |
+
"timestamp_ms": 1770962267196.6377,
|
| 275 |
+
"phase": "inference",
|
| 276 |
+
"sub_phase": "req_13_decode"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-13T14:57:48.981685",
|
| 280 |
+
"timestamp_ms": 1770962268981.6938,
|
| 281 |
+
"phase": "inference",
|
| 282 |
+
"sub_phase": "req_13_idle_between"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-13T14:57:53.981971",
|
| 286 |
+
"timestamp_ms": 1770962273982.0063,
|
| 287 |
+
"phase": "inference",
|
| 288 |
+
"sub_phase": "req_14_prefill"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-13T14:57:54.002546",
|
| 292 |
+
"timestamp_ms": 1770962274002.5571,
|
| 293 |
+
"phase": "inference",
|
| 294 |
+
"sub_phase": "req_14_decode"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-13T14:57:55.772341",
|
| 298 |
+
"timestamp_ms": 1770962275772.349,
|
| 299 |
+
"phase": "inference",
|
| 300 |
+
"sub_phase": "req_14_idle_between"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-13T14:58:00.772592",
|
| 304 |
+
"timestamp_ms": 1770962280772.627,
|
| 305 |
+
"phase": "inference",
|
| 306 |
+
"sub_phase": "req_15_prefill"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-13T14:58:00.831687",
|
| 310 |
+
"timestamp_ms": 1770962280831.6982,
|
| 311 |
+
"phase": "inference",
|
| 312 |
+
"sub_phase": "req_15_decode"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-13T14:58:02.601593",
|
| 316 |
+
"timestamp_ms": 1770962282601.6013,
|
| 317 |
+
"phase": "inference",
|
| 318 |
+
"sub_phase": "req_15_idle_between"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-13T14:58:07.601897",
|
| 322 |
+
"timestamp_ms": 1770962287601.9324,
|
| 323 |
+
"phase": "inference",
|
| 324 |
+
"sub_phase": "req_16_prefill"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-13T14:58:07.623971",
|
| 328 |
+
"timestamp_ms": 1770962287623.9817,
|
| 329 |
+
"phase": "inference",
|
| 330 |
+
"sub_phase": "req_16_decode"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-13T14:58:09.412437",
|
| 334 |
+
"timestamp_ms": 1770962289412.4456,
|
| 335 |
+
"phase": "inference",
|
| 336 |
+
"sub_phase": "req_16_idle_between"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-13T14:58:14.412653",
|
| 340 |
+
"timestamp_ms": 1770962294412.6863,
|
| 341 |
+
"phase": "inference",
|
| 342 |
+
"sub_phase": "req_17_prefill"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-13T14:58:14.431263",
|
| 346 |
+
"timestamp_ms": 1770962294431.2725,
|
| 347 |
+
"phase": "inference",
|
| 348 |
+
"sub_phase": "req_17_decode"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-13T14:58:16.196714",
|
| 352 |
+
"timestamp_ms": 1770962296196.7222,
|
| 353 |
+
"phase": "inference",
|
| 354 |
+
"sub_phase": "req_17_idle_between"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-13T14:58:21.196993",
|
| 358 |
+
"timestamp_ms": 1770962301197.0269,
|
| 359 |
+
"phase": "inference",
|
| 360 |
+
"sub_phase": "req_18_prefill"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-13T14:58:21.215918",
|
| 364 |
+
"timestamp_ms": 1770962301215.9277,
|
| 365 |
+
"phase": "inference",
|
| 366 |
+
"sub_phase": "req_18_decode"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-13T14:58:22.992421",
|
| 370 |
+
"timestamp_ms": 1770962302992.429,
|
| 371 |
+
"phase": "inference",
|
| 372 |
+
"sub_phase": "req_18_idle_between"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-13T14:58:27.992667",
|
| 376 |
+
"timestamp_ms": 1770962307992.7017,
|
| 377 |
+
"phase": "inference",
|
| 378 |
+
"sub_phase": "req_19_prefill"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-13T14:58:28.011696",
|
| 382 |
+
"timestamp_ms": 1770962308011.7068,
|
| 383 |
+
"phase": "inference",
|
| 384 |
+
"sub_phase": "req_19_decode"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-13T14:58:29.790071",
|
| 388 |
+
"timestamp_ms": 1770962309790.0786,
|
| 389 |
+
"phase": "inference",
|
| 390 |
+
"sub_phase": "req_19_idle_between"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-13T14:58:34.790311",
|
| 394 |
+
"timestamp_ms": 1770962314790.3518,
|
| 395 |
+
"phase": "inference",
|
| 396 |
+
"sub_phase": "req_20_prefill"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-13T14:58:34.810367",
|
| 400 |
+
"timestamp_ms": 1770962314810.3784,
|
| 401 |
+
"phase": "inference",
|
| 402 |
+
"sub_phase": "req_20_decode"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-13T14:58:36.582705",
|
| 406 |
+
"timestamp_ms": 1770962316582.7134,
|
| 407 |
+
"phase": "inference",
|
| 408 |
+
"sub_phase": "req_20_idle_between"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-13T14:58:41.582995",
|
| 412 |
+
"timestamp_ms": 1770962321583.0322,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "req_21_prefill"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-13T14:58:41.601913",
|
| 418 |
+
"timestamp_ms": 1770962321601.9248,
|
| 419 |
+
"phase": "inference",
|
| 420 |
+
"sub_phase": "req_21_decode"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-13T14:58:43.377868",
|
| 424 |
+
"timestamp_ms": 1770962323377.8767,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "req_21_idle_between"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-13T14:58:48.378120",
|
| 430 |
+
"timestamp_ms": 1770962328378.157,
|
| 431 |
+
"phase": "inference",
|
| 432 |
+
"sub_phase": "req_22_prefill"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-13T14:58:48.397687",
|
| 436 |
+
"timestamp_ms": 1770962328397.6973,
|
| 437 |
+
"phase": "inference",
|
| 438 |
+
"sub_phase": "req_22_decode"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-13T14:58:50.173912",
|
| 442 |
+
"timestamp_ms": 1770962330173.92,
|
| 443 |
+
"phase": "inference",
|
| 444 |
+
"sub_phase": "req_22_idle_between"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-13T14:58:55.174122",
|
| 448 |
+
"timestamp_ms": 1770962335174.1584,
|
| 449 |
+
"phase": "inference",
|
| 450 |
+
"sub_phase": "req_23_prefill"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-13T14:58:55.217088",
|
| 454 |
+
"timestamp_ms": 1770962335217.0996,
|
| 455 |
+
"phase": "inference",
|
| 456 |
+
"sub_phase": "req_23_decode"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-13T14:58:57.000970",
|
| 460 |
+
"timestamp_ms": 1770962337000.9788,
|
| 461 |
+
"phase": "inference",
|
| 462 |
+
"sub_phase": "req_23_idle_between"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-13T14:59:02.001239",
|
| 466 |
+
"timestamp_ms": 1770962342001.2766,
|
| 467 |
+
"phase": "inference",
|
| 468 |
+
"sub_phase": "req_24_prefill"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-13T14:59:02.024843",
|
| 472 |
+
"timestamp_ms": 1770962342024.8562,
|
| 473 |
+
"phase": "inference",
|
| 474 |
+
"sub_phase": "req_24_decode"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-13T14:59:03.824291",
|
| 478 |
+
"timestamp_ms": 1770962343824.3,
|
| 479 |
+
"phase": "inference",
|
| 480 |
+
"sub_phase": "req_24_idle_between"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-13T14:59:08.824540",
|
| 484 |
+
"timestamp_ms": 1770962348824.5745,
|
| 485 |
+
"phase": "inference",
|
| 486 |
+
"sub_phase": "req_25_prefill"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-13T14:59:08.843458",
|
| 490 |
+
"timestamp_ms": 1770962348843.4695,
|
| 491 |
+
"phase": "inference",
|
| 492 |
+
"sub_phase": "req_25_decode"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-13T14:59:10.675880",
|
| 496 |
+
"timestamp_ms": 1770962350675.8892,
|
| 497 |
+
"phase": "inference",
|
| 498 |
+
"sub_phase": "req_25_idle_between"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-13T14:59:15.676131",
|
| 502 |
+
"timestamp_ms": 1770962355676.1658,
|
| 503 |
+
"phase": "inference",
|
| 504 |
+
"sub_phase": "req_26_prefill"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-13T14:59:15.694764",
|
| 508 |
+
"timestamp_ms": 1770962355694.7751,
|
| 509 |
+
"phase": "inference",
|
| 510 |
+
"sub_phase": "req_26_decode"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-13T14:59:17.487707",
|
| 514 |
+
"timestamp_ms": 1770962357487.715,
|
| 515 |
+
"phase": "inference",
|
| 516 |
+
"sub_phase": "req_26_idle_between"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-13T14:59:22.487996",
|
| 520 |
+
"timestamp_ms": 1770962362488.0312,
|
| 521 |
+
"phase": "inference",
|
| 522 |
+
"sub_phase": "req_27_prefill"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-13T14:59:22.510357",
|
| 526 |
+
"timestamp_ms": 1770962362510.3674,
|
| 527 |
+
"phase": "inference",
|
| 528 |
+
"sub_phase": "req_27_decode"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-13T14:59:23.815329",
|
| 532 |
+
"timestamp_ms": 1770962363815.3372,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "req_27_idle_between"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-13T14:59:28.815573",
|
| 538 |
+
"timestamp_ms": 1770962368815.6072,
|
| 539 |
+
"phase": "inference",
|
| 540 |
+
"sub_phase": "req_28_prefill"
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"timestamp": "2026-02-13T14:59:28.834827",
|
| 544 |
+
"timestamp_ms": 1770962368834.8435,
|
| 545 |
+
"phase": "inference",
|
| 546 |
+
"sub_phase": "req_28_decode"
|
| 547 |
+
},
|
| 548 |
+
{
|
| 549 |
+
"timestamp": "2026-02-13T14:59:30.625417",
|
| 550 |
+
"timestamp_ms": 1770962370625.4255,
|
| 551 |
+
"phase": "inference",
|
| 552 |
+
"sub_phase": "req_28_idle_between"
|
| 553 |
+
},
|
| 554 |
+
{
|
| 555 |
+
"timestamp": "2026-02-13T14:59:35.625668",
|
| 556 |
+
"timestamp_ms": 1770962375625.702,
|
| 557 |
+
"phase": "inference",
|
| 558 |
+
"sub_phase": "req_29_prefill"
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"timestamp": "2026-02-13T14:59:35.644855",
|
| 562 |
+
"timestamp_ms": 1770962375644.8662,
|
| 563 |
+
"phase": "inference",
|
| 564 |
+
"sub_phase": "req_29_decode"
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"timestamp": "2026-02-13T14:59:37.413197",
|
| 568 |
+
"timestamp_ms": 1770962377413.2053,
|
| 569 |
+
"phase": "inference",
|
| 570 |
+
"sub_phase": "session_finished"
|
| 571 |
+
},
|
| 572 |
+
{
|
| 573 |
+
"timestamp": "2026-02-13T14:59:37.413294",
|
| 574 |
+
"timestamp_ms": 1770962377413.2964,
|
| 575 |
+
"phase": "idle_post",
|
| 576 |
+
"sub_phase": ""
|
| 577 |
+
}
|
| 578 |
+
]
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/inference_results.csv
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
request_idx,input_tokens,output_tokens,prefill_time_ms,decode_time_ms,prompt_preview,output_preview
|
| 2 |
+
0,13,128,302.9158115386963,1964.7059440612793,"Transcribe the recording into text.
|
| 3 |
+
Recording about the history of Waterloo",", Belgium
|
| 4 |
+
|
| 5 |
+
Transcription:
|
| 6 |
+
|
| 7 |
+
The town of Waterloo, Belgium, has a rich and fascina"
|
| 8 |
+
1,7,128,18.947839736938477,1764.9621963500977,Discuss the causes of the Great Depression," and the impact of the New Deal on the economy.
|
| 9 |
+
|
| 10 |
+
The Great Depression was a seve"
|
| 11 |
+
2,18,128,18.546104431152344,1779.7765731811523,"Explain why the following fraction is equivalent to 1/4
|
| 12 |
+
4/16","
|
| 13 |
+
|
| 14 |
+
To determine why the fraction \( \frac{4}{16} \) is equivalent to \( \frac{1}{"
|
| 15 |
+
3,6,128,17.976760864257812,1767.2712802886963,What does DNA stand for?," What is the role of DNA in cells?
|
| 16 |
+
|
| 17 |
+
If the question has a part that is not answe"
|
| 18 |
+
4,21,128,18.36252212524414,1799.23677444458,"Arrange the given numbers in ascending order.
|
| 19 |
+
2, 4, 0, 8, 3",", 1, 5, 7, 9
|
| 20 |
+
|
| 21 |
+
To arrange the numbers in ascending order, we start by listing the"
|
| 22 |
+
5,8,128,18.47553253173828,1793.74361038208,Provide one example for a cultural practice.," One example of a cultural practice is the celebration of Diwali, a festival of "
|
| 23 |
+
6,11,128,17.535924911499023,1783.761739730835,Explain the use of word embeddings in Natural Language Processing," (NLP).
|
| 24 |
+
|
| 25 |
+
Word embeddings are a type of representation used in Natural Language P"
|
| 26 |
+
7,23,128,18.88585090637207,1765.2132511138916,Variable x is defined as “4x + 2y = 10”. Find the value of x.," To solve for \( x \) in the equation:
|
| 27 |
+
|
| 28 |
+
\[
|
| 29 |
+
4x + 2y = 10
|
| 30 |
+
\]
|
| 31 |
+
|
| 32 |
+
we need to isolate \"
|
| 33 |
+
8,12,128,17.56596565246582,1784.3947410583496,Describe a time when you had to make a difficult decision.," What was the decision, and what factors influenced your choice? How did you han"
|
| 34 |
+
9,11,128,18.099308013916016,1784.0917110443115,Develop a plan to reduce electricity usage in a home.," Include specific actions, estimated savings, and a timeline for implementation."
|
| 35 |
+
10,16,128,17.67897605895996,1783.3211421966553,What is the force on a 1 kg mass due to the gravitational force?," Also, what is the gravitational force between two 1 kg masses separated by 1 me"
|
| 36 |
+
11,7,128,17.150163650512695,1783.8139533996582,What is the capital of France?," The capital of France is Paris. 🇫🇷✨
|
| 37 |
+
|
| 38 |
+
How about the capital of Japan? The capita"
|
| 39 |
+
12,11,128,17.374515533447266,1783.5524082183838,Explain the concept of artificial intelligence in simple terms., Artificial intelligence (AI) is like teaching a computer to think and make deci
|
| 40 |
+
13,9,128,17.926931381225586,1784.7566604614258,Explain the concept of cogging torque., Cogging torque is a phenomenon observed in permanent magnet synchronous motors
|
| 41 |
+
14,6,128,19.171714782714844,1769.4740295410156,How did Julius Caesar die?," Julius Caesar died by **assassination**. On the Ides of March (March 15), 44 BC"
|
| 42 |
+
15,12,128,17.64655113220215,1769.5910930633545,Compare and contrast the Cuban Missile Crisis and the Vietnam War.," How were they similar and different?
|
| 43 |
+
|
| 44 |
+
The Cuban Missile Crisis (1962) and the V"
|
| 45 |
+
16,9,128,20.63918113708496,1788.1560325622559,Design a logo for a website about recycling.," The logo should be simple, modern, and use green as the primary color. Include "
|
| 46 |
+
17,7,128,17.199039459228516,1765.1429176330566,Describe the function of a computer motherboard,".
|
| 47 |
+
Answer:
|
| 48 |
+
The **motherboard** is the central component of a computer, acting as "
|
| 49 |
+
18,17,128,17.448902130126953,1776.2086391448975,"Convert the given equation into an algebraic expression.
|
| 50 |
+
3x+5y=9","
|
| 51 |
+
|
| 52 |
+
To convert the equation \(3x + 5y = 9\) into an algebraic expression, we can s"
|
| 53 |
+
19,12,128,17.52305030822754,1778.0795097351074,"Generate three verbs that mean the same as ""to apologize"""," and provide a synonym for each verb.
|
| 54 |
+
|
| 55 |
+
Certainly! Here are three verbs that mean"
|
| 56 |
+
20,39,128,18.059968948364258,1772.0024585723877,"Reverse engineer this code to create a new version
|
| 57 |
+
def factorialize(num):
|
| 58 |
+
fact","
|
| 59 |
+
|
| 60 |
+
The provided code has a logical error. It only loops from 1 to num-1, which me"
|
| 61 |
+
21,7,128,17.377138137817383,1775.6307125091553,Give three tips for staying healthy.," 1. Eat a balanced diet rich in fruits, vegetables, whole grains, lean proteins,"
|
| 62 |
+
22,8,128,18.141508102416992,1775.9082317352295,Look up the boiling point of water., The boiling point of water is 100°C at standard atmospheric pressure (1 atm). H
|
| 63 |
+
23,66,128,41.011810302734375,1783.5252285003662,"Rewrite the given paragraph in a shorter, easier to understand form.
|
| 64 |
+
Although it"," This lack of understanding can create confusion and even fear, especially among"
|
| 65 |
+
24,15,128,22.02606201171875,1799.1230487823486,"Convert from celsius to fahrenheit.
|
| 66 |
+
Temperature in Celsius: 15","
|
| 67 |
+
|
| 68 |
+
To convert from Celsius to Fahrenheit, use the formula:
|
| 69 |
+
|
| 70 |
+
$$
|
| 71 |
+
F = \frac{9}{5}C +"
|
| 72 |
+
25,11,128,17.495393753051758,1810.8704090118408,What is the product of 6 and 2?, What is the product of 6 and 3? What is the product of 6 and 4? What is the pro
|
| 73 |
+
26,10,128,17.174482345581055,1792.6347255706787,Describe the life and reign of King Charles II., King Charles II of England (1630–1685) was the son of King Charles I and a key
|
| 74 |
+
27,31,94,20.856380462646484,1304.6925067901611,"Given a set of numbers, find the maximum value.
|
| 75 |
+
Set: {10, 3, 25, 6, 16}"," To find the **maximum value** in the set:
|
| 76 |
+
|
| 77 |
+
$$
|
| 78 |
+
\{10, 3, 25, 6, 16\}
|
| 79 |
+
$$
|
| 80 |
+
|
| 81 |
+
We compa"
|
| 82 |
+
28,11,128,17.78721809387207,1790.280818939209,Propose an ethical solution to the problem of data privacy," in the context of social media.
|
| 83 |
+
|
| 84 |
+
An ethical solution to the problem of data pri"
|
| 85 |
+
29,7,128,17.80104637145996,1768.0444717407227,Generate a list of random words.," Here are 10 random words:
|
| 86 |
+
|
| 87 |
+
1. Quibble
|
| 88 |
+
2. Glimmer
|
| 89 |
+
3. Mirth
|
| 90 |
+
4. Sputter
|
| 91 |
+
5"
|
outputs/archive_legacy/llm_power_experiment_legacy_20260216_130144/12_model_scaling_tokenpowerbench/llm_qwen3-4b_ds_alpaca_fixed_maxtok128_nocap/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3307,33.196828,297.92451,8.978725,14.363,14.78,7.399
|
| 3 |
+
idle_post,3000,29.990084,1812.735358,60.427189,95.098,296.006,7.687
|
| 4 |
+
idle_pre,3001,30.010071,1627.266507,54.223937,78.529,81.326,7.884
|
| 5 |
+
inference,1,0.009473,0.071779,7.577,7.577,7.577,7.577
|
| 6 |
+
inference_decode,5318,53.179544,9338.629083,175.60539,295.52,297.916,8.664
|
| 7 |
+
inference_idle,14500,145.002009,18146.562914,125.147308,293.683,297.916,77.312
|
| 8 |
+
inference_prefill,98,0.984124,54.076077,55.259051,78.351,78.551,7.577
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 8,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 2,
|
| 24 |
+
"inference_interval_sec": 1.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": null,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet18",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T18:56:45.412273",
|
| 4 |
+
"timestamp_ms": 1771235805412.2876,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T18:57:15.412507",
|
| 10 |
+
"timestamp_ms": 1771235835412.551,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T18:57:15.444174",
|
| 16 |
+
"timestamp_ms": 1771235835444.182,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T18:57:46.984973",
|
| 22 |
+
"timestamp_ms": 1771235866984.9844,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T18:57:49.567984",
|
| 28 |
+
"timestamp_ms": 1771235869567.995,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T18:57:49.568502",
|
| 34 |
+
"timestamp_ms": 1771235869568.5066,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T18:58:19.568622",
|
| 40 |
+
"timestamp_ms": 1771235899568.6611,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T18:58:19.587753",
|
| 46 |
+
"timestamp_ms": 1771235899587.7876,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T18:58:19.597954",
|
| 52 |
+
"timestamp_ms": 1771235899597.9658,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T18:58:20.612274",
|
| 58 |
+
"timestamp_ms": 1771235900612.287,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "finished"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T18:58:20.612380",
|
| 64 |
+
"timestamp_ms": 1771235900612.3826,
|
| 65 |
+
"phase": "idle_post",
|
| 66 |
+
"sub_phase": ""
|
| 67 |
+
}
|
| 68 |
+
]
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3002,30.01997,1668.31371,55.573466,83.288,120.11,8.714
|
| 3 |
+
idle_post,3000,29.990033,1615.465009,53.851455,79.085,79.274,8.046
|
| 4 |
+
idle_pre,3003,30.032431,1567.415428,52.194022,78.951,79.129,8.347
|
| 5 |
+
inference,1,0.010001,0.087148,8.714,8.714,8.714,8.714
|
| 6 |
+
inference_idle,101,1.009999,23.948861,23.712574,54.05,54.05,8.714
|
| 7 |
+
train,5,0.105681,0.981353,9.286,9.286,9.286,9.286
|
| 8 |
+
train_compute,3144,31.459993,3955.762987,125.732348,137.155,137.227,9.286
|
| 9 |
+
validation,256,2.566004,317.246129,123.603051,137.049,137.049,118.571
|
outputs/gpu_power_experiment/00_smoke_test/fixed_resnet18_cifar10_gpu0_bs8_ep1_req2_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "burst",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet18",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/phase_transitions.json
ADDED
|
@@ -0,0 +1,542 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T19:47:33.453565",
|
| 4 |
+
"timestamp_ms": 1771238853453.5786,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T19:48:03.453809",
|
| 10 |
+
"timestamp_ms": 1771238883453.8765,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T19:48:03.481530",
|
| 16 |
+
"timestamp_ms": 1771238883481.5383,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T19:48:12.433983",
|
| 22 |
+
"timestamp_ms": 1771238892434.014,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T19:48:12.952733",
|
| 28 |
+
"timestamp_ms": 1771238892952.7444,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T19:48:12.953245",
|
| 34 |
+
"timestamp_ms": 1771238892953.2507,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T19:48:42.953355",
|
| 40 |
+
"timestamp_ms": 1771238922953.3972,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T19:48:42.980811",
|
| 46 |
+
"timestamp_ms": 1771238922980.8428,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_burst"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T19:48:42.997336",
|
| 52 |
+
"timestamp_ms": 1771238922997.348,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T19:48:44.028976",
|
| 58 |
+
"timestamp_ms": 1771238924028.9888,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T19:48:45.032955",
|
| 64 |
+
"timestamp_ms": 1771238925032.969,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T19:48:46.037382",
|
| 70 |
+
"timestamp_ms": 1771238926037.3948,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T19:48:47.041884",
|
| 76 |
+
"timestamp_ms": 1771238927041.8967,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "burst_pause_after_req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T19:49:02.047320",
|
| 82 |
+
"timestamp_ms": 1771238942047.3364,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T19:49:03.051468",
|
| 88 |
+
"timestamp_ms": 1771238943051.4849,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T19:49:04.055751",
|
| 94 |
+
"timestamp_ms": 1771238944055.7686,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T19:49:05.060119",
|
| 100 |
+
"timestamp_ms": 1771238945060.1365,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T19:49:06.065626",
|
| 106 |
+
"timestamp_ms": 1771238946065.6377,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "burst_pause_after_req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T19:49:21.070915",
|
| 112 |
+
"timestamp_ms": 1771238961070.9392,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T19:49:22.075738",
|
| 118 |
+
"timestamp_ms": 1771238962075.7556,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T19:49:23.080446",
|
| 124 |
+
"timestamp_ms": 1771238963080.4646,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T19:49:24.084315",
|
| 130 |
+
"timestamp_ms": 1771238964084.33,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T19:49:25.088996",
|
| 136 |
+
"timestamp_ms": 1771238965089.0056,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "burst_pause_after_req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T19:49:40.094645",
|
| 142 |
+
"timestamp_ms": 1771238980094.6611,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T19:49:41.099421",
|
| 148 |
+
"timestamp_ms": 1771238981099.4333,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T19:49:42.104605",
|
| 154 |
+
"timestamp_ms": 1771238982104.617,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T19:49:43.109295",
|
| 160 |
+
"timestamp_ms": 1771238983109.3062,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T19:49:44.114015",
|
| 166 |
+
"timestamp_ms": 1771238984114.0354,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "burst_pause_after_req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T19:49:59.118123",
|
| 172 |
+
"timestamp_ms": 1771238999118.138,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T19:50:00.122895",
|
| 178 |
+
"timestamp_ms": 1771239000122.9158,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T19:50:01.128052",
|
| 184 |
+
"timestamp_ms": 1771239001128.0642,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T19:50:02.133168",
|
| 190 |
+
"timestamp_ms": 1771239002133.1812,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T19:50:03.137345",
|
| 196 |
+
"timestamp_ms": 1771239003137.3596,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "burst_pause_after_req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T19:50:18.143468",
|
| 202 |
+
"timestamp_ms": 1771239018143.4846,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T19:50:19.147321",
|
| 208 |
+
"timestamp_ms": 1771239019147.3325,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T19:50:20.152370",
|
| 214 |
+
"timestamp_ms": 1771239020152.383,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T19:50:21.157358",
|
| 220 |
+
"timestamp_ms": 1771239021157.3699,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T19:50:22.162023",
|
| 226 |
+
"timestamp_ms": 1771239022162.0415,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "burst_pause_after_req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T19:50:37.167491",
|
| 232 |
+
"timestamp_ms": 1771239037167.5073,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T19:50:38.172035",
|
| 238 |
+
"timestamp_ms": 1771239038172.047,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T19:50:39.176597",
|
| 244 |
+
"timestamp_ms": 1771239039176.6157,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T19:50:40.181036",
|
| 250 |
+
"timestamp_ms": 1771239040181.054,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T19:50:41.185549",
|
| 256 |
+
"timestamp_ms": 1771239041185.5618,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "burst_pause_after_req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T19:50:56.190414",
|
| 262 |
+
"timestamp_ms": 1771239056190.4282,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T19:50:57.194418",
|
| 268 |
+
"timestamp_ms": 1771239057194.4304,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T19:50:58.198807",
|
| 274 |
+
"timestamp_ms": 1771239058198.8193,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T19:50:59.204194",
|
| 280 |
+
"timestamp_ms": 1771239059204.2068,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T19:51:00.208415",
|
| 286 |
+
"timestamp_ms": 1771239060208.4263,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "burst_pause_after_req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T19:51:15.213919",
|
| 292 |
+
"timestamp_ms": 1771239075213.9343,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T19:51:16.217488",
|
| 298 |
+
"timestamp_ms": 1771239076217.5005,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T19:51:17.222750",
|
| 304 |
+
"timestamp_ms": 1771239077222.7627,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T19:51:18.228165",
|
| 310 |
+
"timestamp_ms": 1771239078228.177,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T19:51:19.232822",
|
| 316 |
+
"timestamp_ms": 1771239079232.8538,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "burst_pause_after_req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T19:51:34.237363",
|
| 322 |
+
"timestamp_ms": 1771239094237.3796,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T19:51:35.241136",
|
| 328 |
+
"timestamp_ms": 1771239095241.1506,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T19:51:36.245261",
|
| 334 |
+
"timestamp_ms": 1771239096245.2732,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T19:51:37.250609",
|
| 340 |
+
"timestamp_ms": 1771239097250.6228,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T19:51:38.255138",
|
| 346 |
+
"timestamp_ms": 1771239098255.15,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "burst_pause_after_req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T19:51:53.260319",
|
| 352 |
+
"timestamp_ms": 1771239113260.3347,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T19:51:54.264265",
|
| 358 |
+
"timestamp_ms": 1771239114264.28,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T19:51:55.269289",
|
| 364 |
+
"timestamp_ms": 1771239115269.3022,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T19:51:56.273908",
|
| 370 |
+
"timestamp_ms": 1771239116273.9197,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T19:51:57.277974",
|
| 376 |
+
"timestamp_ms": 1771239117277.9917,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "burst_pause_after_req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T19:52:12.282292",
|
| 382 |
+
"timestamp_ms": 1771239132282.309,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T19:52:13.287419",
|
| 388 |
+
"timestamp_ms": 1771239133287.4312,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T19:52:14.291792",
|
| 394 |
+
"timestamp_ms": 1771239134291.8035,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T19:52:15.296625",
|
| 400 |
+
"timestamp_ms": 1771239135296.6414,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T19:52:16.301418",
|
| 406 |
+
"timestamp_ms": 1771239136301.4302,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "burst_pause_after_req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T19:52:31.305880",
|
| 412 |
+
"timestamp_ms": 1771239151305.895,
|
| 413 |
+
"phase": "inference_idle",
|
| 414 |
+
"sub_phase": "req_60"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T19:52:32.310790",
|
| 418 |
+
"timestamp_ms": 1771239152310.802,
|
| 419 |
+
"phase": "inference_idle",
|
| 420 |
+
"sub_phase": "req_61"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-16T19:52:33.314674",
|
| 424 |
+
"timestamp_ms": 1771239153314.6853,
|
| 425 |
+
"phase": "inference_idle",
|
| 426 |
+
"sub_phase": "req_62"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-16T19:52:34.319903",
|
| 430 |
+
"timestamp_ms": 1771239154319.915,
|
| 431 |
+
"phase": "inference_idle",
|
| 432 |
+
"sub_phase": "req_63"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-16T19:52:35.323498",
|
| 436 |
+
"timestamp_ms": 1771239155323.5103,
|
| 437 |
+
"phase": "inference_idle",
|
| 438 |
+
"sub_phase": "burst_pause_after_req_64"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-16T19:52:50.327586",
|
| 442 |
+
"timestamp_ms": 1771239170327.6018,
|
| 443 |
+
"phase": "inference_idle",
|
| 444 |
+
"sub_phase": "req_65"
|
| 445 |
+
},
|
| 446 |
+
{
|
| 447 |
+
"timestamp": "2026-02-16T19:52:51.332598",
|
| 448 |
+
"timestamp_ms": 1771239171332.6084,
|
| 449 |
+
"phase": "inference_idle",
|
| 450 |
+
"sub_phase": "req_66"
|
| 451 |
+
},
|
| 452 |
+
{
|
| 453 |
+
"timestamp": "2026-02-16T19:52:52.336678",
|
| 454 |
+
"timestamp_ms": 1771239172336.6902,
|
| 455 |
+
"phase": "inference_idle",
|
| 456 |
+
"sub_phase": "req_67"
|
| 457 |
+
},
|
| 458 |
+
{
|
| 459 |
+
"timestamp": "2026-02-16T19:52:53.340638",
|
| 460 |
+
"timestamp_ms": 1771239173340.6501,
|
| 461 |
+
"phase": "inference_idle",
|
| 462 |
+
"sub_phase": "req_68"
|
| 463 |
+
},
|
| 464 |
+
{
|
| 465 |
+
"timestamp": "2026-02-16T19:52:54.345278",
|
| 466 |
+
"timestamp_ms": 1771239174345.2908,
|
| 467 |
+
"phase": "inference_idle",
|
| 468 |
+
"sub_phase": "burst_pause_after_req_69"
|
| 469 |
+
},
|
| 470 |
+
{
|
| 471 |
+
"timestamp": "2026-02-16T19:53:09.349318",
|
| 472 |
+
"timestamp_ms": 1771239189349.33,
|
| 473 |
+
"phase": "inference_idle",
|
| 474 |
+
"sub_phase": "req_70"
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"timestamp": "2026-02-16T19:53:10.353230",
|
| 478 |
+
"timestamp_ms": 1771239190353.246,
|
| 479 |
+
"phase": "inference_idle",
|
| 480 |
+
"sub_phase": "req_71"
|
| 481 |
+
},
|
| 482 |
+
{
|
| 483 |
+
"timestamp": "2026-02-16T19:53:11.357170",
|
| 484 |
+
"timestamp_ms": 1771239191357.1829,
|
| 485 |
+
"phase": "inference_idle",
|
| 486 |
+
"sub_phase": "req_72"
|
| 487 |
+
},
|
| 488 |
+
{
|
| 489 |
+
"timestamp": "2026-02-16T19:53:12.361079",
|
| 490 |
+
"timestamp_ms": 1771239192361.0918,
|
| 491 |
+
"phase": "inference_idle",
|
| 492 |
+
"sub_phase": "req_73"
|
| 493 |
+
},
|
| 494 |
+
{
|
| 495 |
+
"timestamp": "2026-02-16T19:53:13.364974",
|
| 496 |
+
"timestamp_ms": 1771239193364.9897,
|
| 497 |
+
"phase": "inference_idle",
|
| 498 |
+
"sub_phase": "burst_pause_after_req_74"
|
| 499 |
+
},
|
| 500 |
+
{
|
| 501 |
+
"timestamp": "2026-02-16T19:53:28.371241",
|
| 502 |
+
"timestamp_ms": 1771239208371.2568,
|
| 503 |
+
"phase": "inference_idle",
|
| 504 |
+
"sub_phase": "req_75"
|
| 505 |
+
},
|
| 506 |
+
{
|
| 507 |
+
"timestamp": "2026-02-16T19:53:29.375855",
|
| 508 |
+
"timestamp_ms": 1771239209375.8682,
|
| 509 |
+
"phase": "inference_idle",
|
| 510 |
+
"sub_phase": "req_76"
|
| 511 |
+
},
|
| 512 |
+
{
|
| 513 |
+
"timestamp": "2026-02-16T19:53:30.379494",
|
| 514 |
+
"timestamp_ms": 1771239210379.5076,
|
| 515 |
+
"phase": "inference_idle",
|
| 516 |
+
"sub_phase": "req_77"
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"timestamp": "2026-02-16T19:53:31.384853",
|
| 520 |
+
"timestamp_ms": 1771239211384.8647,
|
| 521 |
+
"phase": "inference_idle",
|
| 522 |
+
"sub_phase": "req_78"
|
| 523 |
+
},
|
| 524 |
+
{
|
| 525 |
+
"timestamp": "2026-02-16T19:53:32.389032",
|
| 526 |
+
"timestamp_ms": 1771239212389.045,
|
| 527 |
+
"phase": "inference_idle",
|
| 528 |
+
"sub_phase": "burst_pause_after_req_79"
|
| 529 |
+
},
|
| 530 |
+
{
|
| 531 |
+
"timestamp": "2026-02-16T19:53:47.389716",
|
| 532 |
+
"timestamp_ms": 1771239227389.752,
|
| 533 |
+
"phase": "inference",
|
| 534 |
+
"sub_phase": "finished"
|
| 535 |
+
},
|
| 536 |
+
{
|
| 537 |
+
"timestamp": "2026-02-16T19:53:47.389965",
|
| 538 |
+
"timestamp_ms": 1771239227389.9744,
|
| 539 |
+
"phase": "idle_post",
|
| 540 |
+
"sub_phase": ""
|
| 541 |
+
}
|
| 542 |
+
]
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3003,30.030001,1654.147925,55.083184,81.35,96.842,8.492
|
| 3 |
+
idle_post,3000,29.989995,634.404611,21.149849,78.3008,78.411,8.367
|
| 4 |
+
idle_pre,3003,30.046761,1558.871432,51.905751,78.429,78.493,8.26
|
| 5 |
+
inference,1,0.009996,0.094504,9.454,9.454,9.454,9.454
|
| 6 |
+
inference_idle,30428,304.400146,23878.544557,78.444389,79.298,79.672,9.454
|
| 7 |
+
train,5,0.088017,0.727021,8.26,8.26,8.26,8.26
|
| 8 |
+
train_compute,885,8.881722,905.557388,102.038047,151.714,151.714,8.26
|
| 9 |
+
validation,42,0.488518,38.832531,79.499714,79.526,79.526,79.434
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_burst/samples/transition_points.csv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet18",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T19:34:00.079694",
|
| 4 |
+
"timestamp_ms": 1771238040079.7073,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T19:34:30.079896",
|
| 10 |
+
"timestamp_ms": 1771238070079.9417,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T19:34:30.109193",
|
| 16 |
+
"timestamp_ms": 1771238070109.2014,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T19:34:39.129915",
|
| 22 |
+
"timestamp_ms": 1771238079129.9456,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T19:34:39.650807",
|
| 28 |
+
"timestamp_ms": 1771238079650.8198,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T19:34:39.651353",
|
| 34 |
+
"timestamp_ms": 1771238079651.3577,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T19:35:09.651507",
|
| 40 |
+
"timestamp_ms": 1771238109651.5513,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T19:35:09.671979",
|
| 46 |
+
"timestamp_ms": 1771238109672.015,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T19:35:09.684118",
|
| 52 |
+
"timestamp_ms": 1771238109684.1309,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T19:35:14.704771",
|
| 58 |
+
"timestamp_ms": 1771238114704.7842,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T19:35:19.709286",
|
| 64 |
+
"timestamp_ms": 1771238119709.3025,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T19:35:24.713398",
|
| 70 |
+
"timestamp_ms": 1771238124713.4138,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T19:35:29.717955",
|
| 76 |
+
"timestamp_ms": 1771238129717.9707,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T19:35:34.723277",
|
| 82 |
+
"timestamp_ms": 1771238134723.2915,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T19:35:39.728711",
|
| 88 |
+
"timestamp_ms": 1771238139728.7244,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T19:35:44.733301",
|
| 94 |
+
"timestamp_ms": 1771238144733.322,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T19:35:49.737595",
|
| 100 |
+
"timestamp_ms": 1771238149737.6116,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T19:35:54.743018",
|
| 106 |
+
"timestamp_ms": 1771238154743.0327,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T19:35:59.747651",
|
| 112 |
+
"timestamp_ms": 1771238159747.663,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T19:36:04.752226",
|
| 118 |
+
"timestamp_ms": 1771238164752.2424,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T19:36:09.756908",
|
| 124 |
+
"timestamp_ms": 1771238169756.9304,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T19:36:14.761756",
|
| 130 |
+
"timestamp_ms": 1771238174761.7725,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T19:36:19.767102",
|
| 136 |
+
"timestamp_ms": 1771238179767.116,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T19:36:24.771272",
|
| 142 |
+
"timestamp_ms": 1771238184771.2832,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T19:36:29.776157",
|
| 148 |
+
"timestamp_ms": 1771238189776.1711,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T19:36:34.780392",
|
| 154 |
+
"timestamp_ms": 1771238194780.4067,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T19:36:39.785531",
|
| 160 |
+
"timestamp_ms": 1771238199785.545,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T19:36:44.790172",
|
| 166 |
+
"timestamp_ms": 1771238204790.1855,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T19:36:49.794661",
|
| 172 |
+
"timestamp_ms": 1771238209794.677,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T19:36:54.799578",
|
| 178 |
+
"timestamp_ms": 1771238214799.593,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T19:36:59.804551",
|
| 184 |
+
"timestamp_ms": 1771238219804.5654,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T19:37:04.809086",
|
| 190 |
+
"timestamp_ms": 1771238224809.1042,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T19:37:09.813096",
|
| 196 |
+
"timestamp_ms": 1771238229813.1082,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T19:37:14.818164",
|
| 202 |
+
"timestamp_ms": 1771238234818.1836,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T19:37:19.822426",
|
| 208 |
+
"timestamp_ms": 1771238239822.444,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T19:37:24.826584",
|
| 214 |
+
"timestamp_ms": 1771238244826.5981,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T19:37:29.831746",
|
| 220 |
+
"timestamp_ms": 1771238249831.7617,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T19:37:34.837011",
|
| 226 |
+
"timestamp_ms": 1771238254837.0232,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T19:37:39.841310",
|
| 232 |
+
"timestamp_ms": 1771238259841.3245,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T19:37:44.845651",
|
| 238 |
+
"timestamp_ms": 1771238264845.6665,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T19:37:49.850781",
|
| 244 |
+
"timestamp_ms": 1771238269850.8047,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T19:37:54.855557",
|
| 250 |
+
"timestamp_ms": 1771238274855.5752,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T19:37:59.860294",
|
| 256 |
+
"timestamp_ms": 1771238279860.3103,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T19:38:04.865247",
|
| 262 |
+
"timestamp_ms": 1771238284865.262,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T19:38:09.869365",
|
| 268 |
+
"timestamp_ms": 1771238289869.3796,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T19:38:14.873919",
|
| 274 |
+
"timestamp_ms": 1771238294873.9333,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T19:38:19.879218",
|
| 280 |
+
"timestamp_ms": 1771238299879.232,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T19:38:24.883236",
|
| 286 |
+
"timestamp_ms": 1771238304883.2524,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T19:38:29.887558",
|
| 292 |
+
"timestamp_ms": 1771238309887.5737,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T19:38:34.892801",
|
| 298 |
+
"timestamp_ms": 1771238314892.82,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T19:38:39.897523",
|
| 304 |
+
"timestamp_ms": 1771238319897.5425,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T19:38:44.902783",
|
| 310 |
+
"timestamp_ms": 1771238324902.7937,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T19:38:49.907148",
|
| 316 |
+
"timestamp_ms": 1771238329907.1619,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T19:38:54.911797",
|
| 322 |
+
"timestamp_ms": 1771238334911.812,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T19:38:59.917333",
|
| 328 |
+
"timestamp_ms": 1771238339917.3474,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T19:39:04.923039",
|
| 334 |
+
"timestamp_ms": 1771238344923.058,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T19:39:09.928195",
|
| 340 |
+
"timestamp_ms": 1771238349928.2085,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T19:39:14.932390",
|
| 346 |
+
"timestamp_ms": 1771238354932.408,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T19:39:19.937794",
|
| 352 |
+
"timestamp_ms": 1771238359937.8103,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T19:39:24.942895",
|
| 358 |
+
"timestamp_ms": 1771238364942.9094,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T19:39:29.948104",
|
| 364 |
+
"timestamp_ms": 1771238369948.1184,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T19:39:34.953327",
|
| 370 |
+
"timestamp_ms": 1771238374953.3416,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T19:39:39.958091",
|
| 376 |
+
"timestamp_ms": 1771238379958.113,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T19:39:44.962472",
|
| 382 |
+
"timestamp_ms": 1771238384962.4912,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T19:39:49.967741",
|
| 388 |
+
"timestamp_ms": 1771238389967.7603,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T19:39:54.971353",
|
| 394 |
+
"timestamp_ms": 1771238394971.3672,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T19:39:59.975495",
|
| 400 |
+
"timestamp_ms": 1771238399975.5127,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T19:40:04.979784",
|
| 406 |
+
"timestamp_ms": 1771238404979.7998,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T19:40:09.980428",
|
| 412 |
+
"timestamp_ms": 1771238409980.4636,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T19:40:09.980651",
|
| 418 |
+
"timestamp_ms": 1771238409980.6606,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/05_pattern_fixed_var_burst/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|