Upload batch 16/20 (100 files)
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/config.json +52 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/phase_transitions.json +446 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/phase_power_summary.csv +10 -0
- outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/transition_points.csv +2 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv +13 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_densenet121_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_densenet121_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv +10 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_mobilenetv2_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_mobilenetv2_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv +1 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv +1 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv +0 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv +9 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv +13 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json +52 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json +422 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png +3 -0
- outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html +0 -0
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": true,
|
| 35 |
+
"ramp_start_w": 360,
|
| 36 |
+
"ramp_end_w": 600,
|
| 37 |
+
"ramp_steps": 10,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet18",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,446 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T22:03:26.050618",
|
| 4 |
+
"timestamp_ms": 1771247006050.6316,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T22:03:56.051116",
|
| 10 |
+
"timestamp_ms": 1771247036051.1575,
|
| 11 |
+
"phase": "ramp",
|
| 12 |
+
"sub_phase": "up_360W_to_600W"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T22:04:26.697239",
|
| 16 |
+
"timestamp_ms": 1771247066697.275,
|
| 17 |
+
"phase": "ramp",
|
| 18 |
+
"sub_phase": "up_complete"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T22:04:26.697337",
|
| 22 |
+
"timestamp_ms": 1771247066697.344,
|
| 23 |
+
"phase": "train",
|
| 24 |
+
"sub_phase": "start"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T22:04:26.729365",
|
| 28 |
+
"timestamp_ms": 1771247066729.3723,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "epoch_1_start"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T22:04:36.021598",
|
| 34 |
+
"timestamp_ms": 1771247076021.63,
|
| 35 |
+
"phase": "validation",
|
| 36 |
+
"sub_phase": "epoch_1"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T22:04:36.612821",
|
| 40 |
+
"timestamp_ms": 1771247076612.84,
|
| 41 |
+
"phase": "train",
|
| 42 |
+
"sub_phase": "finished"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T22:04:36.613331",
|
| 46 |
+
"timestamp_ms": 1771247076613.3354,
|
| 47 |
+
"phase": "idle_mid",
|
| 48 |
+
"sub_phase": ""
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T22:05:06.613439",
|
| 52 |
+
"timestamp_ms": 1771247106613.4807,
|
| 53 |
+
"phase": "inference",
|
| 54 |
+
"sub_phase": "start"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T22:05:06.632958",
|
| 58 |
+
"timestamp_ms": 1771247106632.972,
|
| 59 |
+
"phase": "inference",
|
| 60 |
+
"sub_phase": "start_pattern_fixed"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T22:05:06.646319",
|
| 64 |
+
"timestamp_ms": 1771247106646.3325,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_0"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T22:05:11.666809",
|
| 70 |
+
"timestamp_ms": 1771247111666.8237,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_1"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T22:05:16.671760",
|
| 76 |
+
"timestamp_ms": 1771247116671.7822,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_2"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T22:05:21.676016",
|
| 82 |
+
"timestamp_ms": 1771247121676.0308,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_3"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T22:05:26.680227",
|
| 88 |
+
"timestamp_ms": 1771247126680.2432,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_4"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T22:05:31.684260",
|
| 94 |
+
"timestamp_ms": 1771247131684.2769,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_5"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T22:05:36.690000",
|
| 100 |
+
"timestamp_ms": 1771247136690.0117,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_6"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T22:05:41.695444",
|
| 106 |
+
"timestamp_ms": 1771247141695.4585,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_7"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T22:05:46.699846",
|
| 112 |
+
"timestamp_ms": 1771247146699.862,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_8"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T22:05:51.705351",
|
| 118 |
+
"timestamp_ms": 1771247151705.3672,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_9"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T22:05:56.709613",
|
| 124 |
+
"timestamp_ms": 1771247156709.628,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_10"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T22:06:01.713796",
|
| 130 |
+
"timestamp_ms": 1771247161713.8127,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_11"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T22:06:06.717719",
|
| 136 |
+
"timestamp_ms": 1771247166717.7324,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_12"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T22:06:11.723433",
|
| 142 |
+
"timestamp_ms": 1771247171723.4485,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_13"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T22:06:16.728818",
|
| 148 |
+
"timestamp_ms": 1771247176728.8403,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_14"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T22:06:21.734912",
|
| 154 |
+
"timestamp_ms": 1771247181734.926,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_15"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T22:06:26.739078",
|
| 160 |
+
"timestamp_ms": 1771247186739.0938,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_16"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T22:06:31.743942",
|
| 166 |
+
"timestamp_ms": 1771247191743.9568,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_17"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T22:06:36.748167",
|
| 172 |
+
"timestamp_ms": 1771247196748.182,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_18"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T22:06:41.753390",
|
| 178 |
+
"timestamp_ms": 1771247201753.4058,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_19"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T22:06:46.757684",
|
| 184 |
+
"timestamp_ms": 1771247206757.703,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_20"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T22:06:51.761934",
|
| 190 |
+
"timestamp_ms": 1771247211761.952,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_21"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T22:06:56.766669",
|
| 196 |
+
"timestamp_ms": 1771247216766.6836,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_22"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T22:07:01.771529",
|
| 202 |
+
"timestamp_ms": 1771247221771.544,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_23"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T22:07:06.776585",
|
| 208 |
+
"timestamp_ms": 1771247226776.5994,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_24"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T22:07:11.780682",
|
| 214 |
+
"timestamp_ms": 1771247231780.6982,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_25"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T22:07:16.785398",
|
| 220 |
+
"timestamp_ms": 1771247236785.419,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_26"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T22:07:21.790235",
|
| 226 |
+
"timestamp_ms": 1771247241790.251,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_27"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T22:07:26.795200",
|
| 232 |
+
"timestamp_ms": 1771247246795.2217,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_28"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T22:07:31.799022",
|
| 238 |
+
"timestamp_ms": 1771247251799.0369,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_29"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T22:07:36.803504",
|
| 244 |
+
"timestamp_ms": 1771247256803.5232,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_30"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T22:07:41.807371",
|
| 250 |
+
"timestamp_ms": 1771247261807.3845,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_31"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T22:07:46.812661",
|
| 256 |
+
"timestamp_ms": 1771247266812.68,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_32"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T22:07:51.817822",
|
| 262 |
+
"timestamp_ms": 1771247271817.8406,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_33"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T22:07:56.822313",
|
| 268 |
+
"timestamp_ms": 1771247276822.3298,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_34"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T22:08:01.827447",
|
| 274 |
+
"timestamp_ms": 1771247281827.4607,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_35"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T22:08:06.832628",
|
| 280 |
+
"timestamp_ms": 1771247286832.649,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_36"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T22:08:11.838021",
|
| 286 |
+
"timestamp_ms": 1771247291838.0354,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_37"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T22:08:16.843889",
|
| 292 |
+
"timestamp_ms": 1771247296843.902,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_38"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T22:08:21.847956",
|
| 298 |
+
"timestamp_ms": 1771247301847.972,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_39"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T22:08:26.853003",
|
| 304 |
+
"timestamp_ms": 1771247306853.0164,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_40"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T22:08:31.857680",
|
| 310 |
+
"timestamp_ms": 1771247311857.6921,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_41"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T22:08:36.863194",
|
| 316 |
+
"timestamp_ms": 1771247316863.2068,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_42"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T22:08:41.867330",
|
| 322 |
+
"timestamp_ms": 1771247321867.344,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_43"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T22:08:46.871606",
|
| 328 |
+
"timestamp_ms": 1771247326871.6208,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_44"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T22:08:51.875692",
|
| 334 |
+
"timestamp_ms": 1771247331875.7068,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_45"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T22:08:56.880644",
|
| 340 |
+
"timestamp_ms": 1771247336880.6626,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_46"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T22:09:01.885012",
|
| 346 |
+
"timestamp_ms": 1771247341885.0254,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_47"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T22:09:06.888794",
|
| 352 |
+
"timestamp_ms": 1771247346888.806,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_48"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T22:09:11.892826",
|
| 358 |
+
"timestamp_ms": 1771247351892.8496,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_49"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T22:09:16.896843",
|
| 364 |
+
"timestamp_ms": 1771247356896.857,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_50"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T22:09:21.902097",
|
| 370 |
+
"timestamp_ms": 1771247361902.1113,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_51"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T22:09:26.906577",
|
| 376 |
+
"timestamp_ms": 1771247366906.5923,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_52"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T22:09:31.911862",
|
| 382 |
+
"timestamp_ms": 1771247371911.8733,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_53"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T22:09:36.915958",
|
| 388 |
+
"timestamp_ms": 1771247376915.9707,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_54"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T22:09:41.920338",
|
| 394 |
+
"timestamp_ms": 1771247381920.3528,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_55"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T22:09:46.924879",
|
| 400 |
+
"timestamp_ms": 1771247386924.8938,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_56"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T22:09:51.928851",
|
| 406 |
+
"timestamp_ms": 1771247391928.8645,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_57"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T22:09:56.933675",
|
| 412 |
+
"timestamp_ms": 1771247396933.6907,
|
| 413 |
+
"phase": "inference_idle",
|
| 414 |
+
"sub_phase": "req_58"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T22:10:01.937887",
|
| 418 |
+
"timestamp_ms": 1771247401937.8997,
|
| 419 |
+
"phase": "inference_idle",
|
| 420 |
+
"sub_phase": "req_59"
|
| 421 |
+
},
|
| 422 |
+
{
|
| 423 |
+
"timestamp": "2026-02-16T22:10:06.938486",
|
| 424 |
+
"timestamp_ms": 1771247406938.5193,
|
| 425 |
+
"phase": "inference",
|
| 426 |
+
"sub_phase": "finished"
|
| 427 |
+
},
|
| 428 |
+
{
|
| 429 |
+
"timestamp": "2026-02-16T22:10:06.938721",
|
| 430 |
+
"timestamp_ms": 1771247406938.731,
|
| 431 |
+
"phase": "ramp",
|
| 432 |
+
"sub_phase": "down_600W_to_360W"
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"timestamp": "2026-02-16T22:10:37.391273",
|
| 436 |
+
"timestamp_ms": 1771247437391.3074,
|
| 437 |
+
"phase": "ramp",
|
| 438 |
+
"sub_phase": "down_complete"
|
| 439 |
+
},
|
| 440 |
+
{
|
| 441 |
+
"timestamp": "2026-02-16T22:10:37.391375",
|
| 442 |
+
"timestamp_ms": 1771247437391.382,
|
| 443 |
+
"phase": "idle_post",
|
| 444 |
+
"sub_phase": ""
|
| 445 |
+
}
|
| 446 |
+
]
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3002,30.02002,1672.458805,55.711518,79.12,98.075,7.486
|
| 3 |
+
idle_post,3000,29.989998,289.885517,9.665801,14.924,15.149,8.061
|
| 4 |
+
idle_pre,3001,30.010069,1622.11828,54.052498,78.315,78.612,7.987
|
| 5 |
+
inference,1,0.009946,0.090863,9.136,9.136,9.136,9.136
|
| 6 |
+
inference_idle,30017,300.288878,23389.149197,77.888607,78.593,78.941,9.136
|
| 7 |
+
ramp,6112,61.130831,1577.35123,25.805848,78.221,78.47,7.923
|
| 8 |
+
train,6,0.112894,2.411269,24.3335,27.505,27.505,8.476
|
| 9 |
+
train_compute,918,9.209901,973.463658,105.785074,148.302,148.302,27.505
|
| 10 |
+
validation,51,0.567308,44.945391,79.266118,79.403,79.403,78.866
|
outputs/gpu_power_experiment/04_control_ramp/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_ramp_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
| 2 |
+
6112,61193.613,train_compute,train_compute,27.505,80.904,53.399,10.034664,5321.453677,up
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "convnext_tiny",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:36:19.154266",
|
| 4 |
+
"timestamp_ms": 1771241779154.2883,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:36:49.154574",
|
| 10 |
+
"timestamp_ms": 1771241809154.6125,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:36:49.185157",
|
| 16 |
+
"timestamp_ms": 1771241809185.164,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:37:30.499125",
|
| 22 |
+
"timestamp_ms": 1771241850499.1355,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:37:32.916963",
|
| 28 |
+
"timestamp_ms": 1771241852916.9749,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:37:32.918699",
|
| 34 |
+
"timestamp_ms": 1771241852918.705,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:38:02.918895",
|
| 40 |
+
"timestamp_ms": 1771241882918.9375,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:38:03.081209",
|
| 46 |
+
"timestamp_ms": 1771241883081.2512,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:38:03.109793",
|
| 52 |
+
"timestamp_ms": 1771241883109.806,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:38:08.269370",
|
| 58 |
+
"timestamp_ms": 1771241888269.3948,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:38:13.279451",
|
| 64 |
+
"timestamp_ms": 1771241893279.464,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:38:18.289190",
|
| 70 |
+
"timestamp_ms": 1771241898289.2026,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:38:23.299172",
|
| 76 |
+
"timestamp_ms": 1771241903299.1848,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:38:28.309072",
|
| 82 |
+
"timestamp_ms": 1771241908309.0852,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:38:33.320141",
|
| 88 |
+
"timestamp_ms": 1771241913320.161,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:38:38.330077",
|
| 94 |
+
"timestamp_ms": 1771241918330.0896,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:38:43.340062",
|
| 100 |
+
"timestamp_ms": 1771241923340.0747,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:38:48.350138",
|
| 106 |
+
"timestamp_ms": 1771241928350.153,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:38:53.360304",
|
| 112 |
+
"timestamp_ms": 1771241933360.316,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:38:58.370840",
|
| 118 |
+
"timestamp_ms": 1771241938370.855,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:39:03.380730",
|
| 124 |
+
"timestamp_ms": 1771241943380.7422,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:39:08.390508",
|
| 130 |
+
"timestamp_ms": 1771241948390.5198,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:39:13.400865",
|
| 136 |
+
"timestamp_ms": 1771241953400.8804,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:39:18.411226",
|
| 142 |
+
"timestamp_ms": 1771241958411.2385,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:39:23.421276",
|
| 148 |
+
"timestamp_ms": 1771241963421.2913,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:39:28.431296",
|
| 154 |
+
"timestamp_ms": 1771241968431.3123,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:39:33.441252",
|
| 160 |
+
"timestamp_ms": 1771241973441.2634,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:39:38.451126",
|
| 166 |
+
"timestamp_ms": 1771241978451.1372,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:39:43.461182",
|
| 172 |
+
"timestamp_ms": 1771241983461.1968,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:39:48.471233",
|
| 178 |
+
"timestamp_ms": 1771241988471.2478,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:39:53.481301",
|
| 184 |
+
"timestamp_ms": 1771241993481.313,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:39:58.490891",
|
| 190 |
+
"timestamp_ms": 1771241998490.9023,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:40:03.500759",
|
| 196 |
+
"timestamp_ms": 1771242003500.7712,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:40:08.510882",
|
| 202 |
+
"timestamp_ms": 1771242008510.8982,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:40:13.521135",
|
| 208 |
+
"timestamp_ms": 1771242013521.1477,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:40:18.531006",
|
| 214 |
+
"timestamp_ms": 1771242018531.0183,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:40:23.540950",
|
| 220 |
+
"timestamp_ms": 1771242023540.9639,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:40:28.551105",
|
| 226 |
+
"timestamp_ms": 1771242028551.1196,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:40:33.561003",
|
| 232 |
+
"timestamp_ms": 1771242033561.0183,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:40:38.571386",
|
| 238 |
+
"timestamp_ms": 1771242038571.3984,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:40:43.581282",
|
| 244 |
+
"timestamp_ms": 1771242043581.2947,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:40:48.591341",
|
| 250 |
+
"timestamp_ms": 1771242048591.3535,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:40:53.601300",
|
| 256 |
+
"timestamp_ms": 1771242053601.313,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:40:58.611510",
|
| 262 |
+
"timestamp_ms": 1771242058611.5234,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:41:03.621535",
|
| 268 |
+
"timestamp_ms": 1771242063621.55,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:41:08.631639",
|
| 274 |
+
"timestamp_ms": 1771242068631.655,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:41:13.641810",
|
| 280 |
+
"timestamp_ms": 1771242073641.8252,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:41:18.651674",
|
| 286 |
+
"timestamp_ms": 1771242078651.6865,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:41:23.661514",
|
| 292 |
+
"timestamp_ms": 1771242083661.5256,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:41:28.671416",
|
| 298 |
+
"timestamp_ms": 1771242088671.4329,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:41:33.681734",
|
| 304 |
+
"timestamp_ms": 1771242093681.7466,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:41:38.691679",
|
| 310 |
+
"timestamp_ms": 1771242098691.6921,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:41:43.701261",
|
| 316 |
+
"timestamp_ms": 1771242103701.2751,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:41:48.711301",
|
| 322 |
+
"timestamp_ms": 1771242108711.3125,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:41:53.721941",
|
| 328 |
+
"timestamp_ms": 1771242113721.9534,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:41:58.731927",
|
| 334 |
+
"timestamp_ms": 1771242118731.9397,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:42:03.742088",
|
| 340 |
+
"timestamp_ms": 1771242123742.101,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:42:08.751904",
|
| 346 |
+
"timestamp_ms": 1771242128751.9167,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:42:13.762367",
|
| 352 |
+
"timestamp_ms": 1771242133762.3823,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:42:18.772359",
|
| 358 |
+
"timestamp_ms": 1771242138772.3713,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:42:23.782295",
|
| 364 |
+
"timestamp_ms": 1771242143782.3071,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:42:28.792239",
|
| 370 |
+
"timestamp_ms": 1771242148792.2517,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:42:33.801782",
|
| 376 |
+
"timestamp_ms": 1771242153801.7942,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:42:38.811693",
|
| 382 |
+
"timestamp_ms": 1771242158811.705,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:42:43.821681",
|
| 388 |
+
"timestamp_ms": 1771242163821.6965,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:42:48.831761",
|
| 394 |
+
"timestamp_ms": 1771242168831.7766,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:42:53.841839",
|
| 400 |
+
"timestamp_ms": 1771242173841.8516,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:42:58.851770",
|
| 406 |
+
"timestamp_ms": 1771242178851.782,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:43:03.852405",
|
| 412 |
+
"timestamp_ms": 1771242183852.4412,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:43:03.852647",
|
| 418 |
+
"timestamp_ms": 1771242183852.656,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3015,30.152742,2000.125264,66.338335,153.92,392.004,8.381
|
| 3 |
+
idle_post,3000,29.990089,1275.224627,42.510264,78.1774,78.729,7.991
|
| 4 |
+
idle_pre,3004,30.065334,1611.782025,53.647422,78.654,78.958,8.252
|
| 5 |
+
inference,4,0.037277,0.353495,9.483,9.483,9.483,9.483
|
| 6 |
+
inference_idle,30064,300.743112,23803.24612,79.14747,82.662,83.41,9.483
|
| 7 |
+
train,31,0.338168,4.322647,13.168935,14.349,14.349,8.252
|
| 8 |
+
train_compute,4095,40.976584,13337.912043,325.64016,512.333,514.439,14.349
|
| 9 |
+
validation,232,2.393998,622.287107,264.82706,396.614,396.614,89.105
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_convnext_tiny_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
| 2 |
+
3909,39152.817,train_compute,train_compute,122.692,176.249,53.557,10.036373,5336.290287,up
|
| 3 |
+
4059,40652.778,train_compute,train_compute,151.678,242.084,90.406,9.993118,9046.826025,up
|
| 4 |
+
4109,41152.788,train_compute,train_compute,242.084,428.613,186.529,8.546315,21825.664,up
|
| 5 |
+
4159,41652.783,train_compute,train_compute,428.613,505.0,76.387,9.998486,7639.856653,up
|
| 6 |
+
6309,63152.783,train_compute,train_compute,512.539,403.417,-109.122,9.9655,-10949.977445,down
|
| 7 |
+
6359,63652.78,train_compute,train_compute,403.417,193.493,-209.924,9.963141,-21070.062024,down
|
| 8 |
+
6409,64152.782,train_compute,train_compute,193.493,105.942,-87.551,9.964927,-8785.914854,down
|
| 9 |
+
7199,72154.625,validation,validation,89.105,229.929,140.824,9.012308,15625.741888,up
|
| 10 |
+
7249,72654.728,validation,validation,229.929,376.371,146.442,10.116682,14475.299257,up
|
| 11 |
+
7399,74154.656,idle_mid,idle_mid,392.004,310.8,-81.204,10.045509,-8083.612313,down
|
| 12 |
+
7449,74654.655,idle_mid,idle_mid,310.8,159.682,-151.118,10.044128,-15045.407722,down
|
| 13 |
+
7599,76154.612,idle_mid,idle_mid,153.92,93.475,-60.445,9.970711,-6062.255693,down
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_densenet121_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_densenet121_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "densenet121",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_densenet121_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:15:20.275582",
|
| 4 |
+
"timestamp_ms": 1771240520275.5962,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:15:50.275739",
|
| 10 |
+
"timestamp_ms": 1771240550275.7808,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:15:50.309752",
|
| 16 |
+
"timestamp_ms": 1771240550309.76,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:16:28.517082",
|
| 22 |
+
"timestamp_ms": 1771240588517.0933,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:16:30.943196",
|
| 28 |
+
"timestamp_ms": 1771240590943.2078,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:16:30.944908",
|
| 34 |
+
"timestamp_ms": 1771240590944.9153,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:17:00.945066",
|
| 40 |
+
"timestamp_ms": 1771240620945.1086,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:17:01.066818",
|
| 46 |
+
"timestamp_ms": 1771240621066.8442,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:17:01.111253",
|
| 52 |
+
"timestamp_ms": 1771240621111.277,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:17:06.256206",
|
| 58 |
+
"timestamp_ms": 1771240626256.2354,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:17:11.268376",
|
| 64 |
+
"timestamp_ms": 1771240631268.3918,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:17:16.280686",
|
| 70 |
+
"timestamp_ms": 1771240636280.6985,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:17:21.294705",
|
| 76 |
+
"timestamp_ms": 1771240641294.7249,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:17:26.307371",
|
| 82 |
+
"timestamp_ms": 1771240646307.3867,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:17:31.319125",
|
| 88 |
+
"timestamp_ms": 1771240651319.1382,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:17:36.331292",
|
| 94 |
+
"timestamp_ms": 1771240656331.3062,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:17:41.343068",
|
| 100 |
+
"timestamp_ms": 1771240661343.081,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:17:46.355481",
|
| 106 |
+
"timestamp_ms": 1771240666355.4934,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:17:51.367633",
|
| 112 |
+
"timestamp_ms": 1771240671367.6448,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:17:56.379285",
|
| 118 |
+
"timestamp_ms": 1771240676379.299,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:18:01.391148",
|
| 124 |
+
"timestamp_ms": 1771240681391.1614,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:18:06.403370",
|
| 130 |
+
"timestamp_ms": 1771240686403.383,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:18:11.415587",
|
| 136 |
+
"timestamp_ms": 1771240691415.6,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:18:16.428103",
|
| 142 |
+
"timestamp_ms": 1771240696428.1152,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:18:21.440257",
|
| 148 |
+
"timestamp_ms": 1771240701440.2695,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:18:26.452296",
|
| 154 |
+
"timestamp_ms": 1771240706452.309,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:18:31.464753",
|
| 160 |
+
"timestamp_ms": 1771240711464.769,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:18:36.477247",
|
| 166 |
+
"timestamp_ms": 1771240716477.2585,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:18:41.489365",
|
| 172 |
+
"timestamp_ms": 1771240721489.3818,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:18:46.501424",
|
| 178 |
+
"timestamp_ms": 1771240726501.4365,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:18:51.513469",
|
| 184 |
+
"timestamp_ms": 1771240731513.4841,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:18:56.525266",
|
| 190 |
+
"timestamp_ms": 1771240736525.2788,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:19:01.537376",
|
| 196 |
+
"timestamp_ms": 1771240741537.3914,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:19:06.671866",
|
| 202 |
+
"timestamp_ms": 1771240746671.8813,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:19:11.684159",
|
| 208 |
+
"timestamp_ms": 1771240751684.1716,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:19:16.696173",
|
| 214 |
+
"timestamp_ms": 1771240756696.185,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:19:21.708144",
|
| 220 |
+
"timestamp_ms": 1771240761708.1594,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:19:26.720407",
|
| 226 |
+
"timestamp_ms": 1771240766720.4214,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:19:31.732482",
|
| 232 |
+
"timestamp_ms": 1771240771732.4978,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:19:36.744174",
|
| 238 |
+
"timestamp_ms": 1771240776744.1865,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:19:41.756449",
|
| 244 |
+
"timestamp_ms": 1771240781756.4646,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:19:46.768905",
|
| 250 |
+
"timestamp_ms": 1771240786768.9207,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:19:51.780992",
|
| 256 |
+
"timestamp_ms": 1771240791781.0051,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:19:56.793276",
|
| 262 |
+
"timestamp_ms": 1771240796793.289,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:20:01.805448",
|
| 268 |
+
"timestamp_ms": 1771240801805.4612,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:20:06.817554",
|
| 274 |
+
"timestamp_ms": 1771240806817.5693,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:20:11.829945",
|
| 280 |
+
"timestamp_ms": 1771240811829.9612,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:20:16.841837",
|
| 286 |
+
"timestamp_ms": 1771240816841.8499,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:20:21.853943",
|
| 292 |
+
"timestamp_ms": 1771240821853.9546,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:20:26.865789",
|
| 298 |
+
"timestamp_ms": 1771240826865.8032,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:20:31.877578",
|
| 304 |
+
"timestamp_ms": 1771240831877.5916,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:20:36.890056",
|
| 310 |
+
"timestamp_ms": 1771240836890.0688,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:20:41.902144",
|
| 316 |
+
"timestamp_ms": 1771240841902.1562,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:20:46.914376",
|
| 322 |
+
"timestamp_ms": 1771240846914.3923,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:20:51.926321",
|
| 328 |
+
"timestamp_ms": 1771240851926.3333,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:20:56.938794",
|
| 334 |
+
"timestamp_ms": 1771240856938.8093,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:21:01.950663",
|
| 340 |
+
"timestamp_ms": 1771240861950.6755,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:21:06.962996",
|
| 346 |
+
"timestamp_ms": 1771240866963.01,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:21:11.975388",
|
| 352 |
+
"timestamp_ms": 1771240871975.4011,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:21:16.987660",
|
| 358 |
+
"timestamp_ms": 1771240876987.673,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:21:21.999678",
|
| 364 |
+
"timestamp_ms": 1771240881999.69,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:21:27.011589",
|
| 370 |
+
"timestamp_ms": 1771240887011.6013,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:21:32.024061",
|
| 376 |
+
"timestamp_ms": 1771240892024.0781,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:21:37.036090",
|
| 382 |
+
"timestamp_ms": 1771240897036.1052,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:21:42.048334",
|
| 388 |
+
"timestamp_ms": 1771240902048.347,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:21:47.059814",
|
| 394 |
+
"timestamp_ms": 1771240907059.8267,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:21:52.080783",
|
| 400 |
+
"timestamp_ms": 1771240912080.803,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:21:57.093054",
|
| 406 |
+
"timestamp_ms": 1771240917093.0696,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:22:02.093620",
|
| 412 |
+
"timestamp_ms": 1771240922093.6453,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:22:02.093770",
|
| 418 |
+
"timestamp_ms": 1771240922093.7754,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "efficientnet_b0",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:58:03.458787",
|
| 4 |
+
"timestamp_ms": 1771243083458.8,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:58:33.458993",
|
| 10 |
+
"timestamp_ms": 1771243113459.0327,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:58:33.488060",
|
| 16 |
+
"timestamp_ms": 1771243113488.0674,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:59:03.490206",
|
| 22 |
+
"timestamp_ms": 1771243143490.2166,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:59:05.684624",
|
| 28 |
+
"timestamp_ms": 1771243145684.6362,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:59:05.685785",
|
| 34 |
+
"timestamp_ms": 1771243145685.7915,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:59:35.685953",
|
| 40 |
+
"timestamp_ms": 1771243175685.994,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:59:35.804629",
|
| 46 |
+
"timestamp_ms": 1771243175804.6658,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:59:35.827233",
|
| 52 |
+
"timestamp_ms": 1771243175827.2502,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:59:41.043080",
|
| 58 |
+
"timestamp_ms": 1771243181043.096,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:59:46.051553",
|
| 64 |
+
"timestamp_ms": 1771243186051.5679,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:59:51.059641",
|
| 70 |
+
"timestamp_ms": 1771243191059.654,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:59:56.068058",
|
| 76 |
+
"timestamp_ms": 1771243196068.073,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T21:00:01.076735",
|
| 82 |
+
"timestamp_ms": 1771243201076.754,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T21:00:06.085099",
|
| 88 |
+
"timestamp_ms": 1771243206085.1118,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T21:00:11.093038",
|
| 94 |
+
"timestamp_ms": 1771243211093.051,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T21:00:16.100855",
|
| 100 |
+
"timestamp_ms": 1771243216100.8674,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T21:00:21.108312",
|
| 106 |
+
"timestamp_ms": 1771243221108.3242,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T21:00:26.115906",
|
| 112 |
+
"timestamp_ms": 1771243226115.922,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T21:00:31.124553",
|
| 118 |
+
"timestamp_ms": 1771243231124.5654,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T21:00:36.132397",
|
| 124 |
+
"timestamp_ms": 1771243236132.4087,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T21:00:41.140023",
|
| 130 |
+
"timestamp_ms": 1771243241140.0388,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T21:00:46.148206",
|
| 136 |
+
"timestamp_ms": 1771243246148.2207,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T21:00:51.154641",
|
| 142 |
+
"timestamp_ms": 1771243251154.6553,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T21:00:56.162765",
|
| 148 |
+
"timestamp_ms": 1771243256162.7764,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T21:01:01.171338",
|
| 154 |
+
"timestamp_ms": 1771243261171.351,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T21:01:06.179056",
|
| 160 |
+
"timestamp_ms": 1771243266179.0698,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T21:01:11.187247",
|
| 166 |
+
"timestamp_ms": 1771243271187.259,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T21:01:16.195158",
|
| 172 |
+
"timestamp_ms": 1771243276195.1704,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T21:01:21.206989",
|
| 178 |
+
"timestamp_ms": 1771243281207.0034,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T21:01:26.215301",
|
| 184 |
+
"timestamp_ms": 1771243286215.3135,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T21:01:31.223140",
|
| 190 |
+
"timestamp_ms": 1771243291223.1526,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T21:01:36.231667",
|
| 196 |
+
"timestamp_ms": 1771243296231.6821,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T21:01:41.239026",
|
| 202 |
+
"timestamp_ms": 1771243301239.038,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T21:01:46.247293",
|
| 208 |
+
"timestamp_ms": 1771243306247.3088,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T21:01:51.255376",
|
| 214 |
+
"timestamp_ms": 1771243311255.389,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T21:01:56.263810",
|
| 220 |
+
"timestamp_ms": 1771243316263.8228,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T21:02:01.275277",
|
| 226 |
+
"timestamp_ms": 1771243321275.2922,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T21:02:06.283004",
|
| 232 |
+
"timestamp_ms": 1771243326283.0168,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T21:02:11.290823",
|
| 238 |
+
"timestamp_ms": 1771243331290.8408,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T21:02:16.298894",
|
| 244 |
+
"timestamp_ms": 1771243336298.9062,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T21:02:21.308308",
|
| 250 |
+
"timestamp_ms": 1771243341308.3284,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T21:02:26.315943",
|
| 256 |
+
"timestamp_ms": 1771243346315.9546,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T21:02:31.324257",
|
| 262 |
+
"timestamp_ms": 1771243351324.269,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T21:02:36.332632",
|
| 268 |
+
"timestamp_ms": 1771243356332.6445,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T21:02:41.344277",
|
| 274 |
+
"timestamp_ms": 1771243361344.2925,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T21:02:46.352373",
|
| 280 |
+
"timestamp_ms": 1771243366352.3855,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T21:02:51.360458",
|
| 286 |
+
"timestamp_ms": 1771243371360.4731,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T21:02:56.368388",
|
| 292 |
+
"timestamp_ms": 1771243376368.4033,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T21:03:01.376720",
|
| 298 |
+
"timestamp_ms": 1771243381376.7317,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T21:03:06.384810",
|
| 304 |
+
"timestamp_ms": 1771243386384.8215,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T21:03:11.392939",
|
| 310 |
+
"timestamp_ms": 1771243391392.951,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T21:03:16.403373",
|
| 316 |
+
"timestamp_ms": 1771243396403.3887,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T21:03:21.411567",
|
| 322 |
+
"timestamp_ms": 1771243401411.5833,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T21:03:26.419477",
|
| 328 |
+
"timestamp_ms": 1771243406419.492,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T21:03:31.427556",
|
| 334 |
+
"timestamp_ms": 1771243411427.5684,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T21:03:36.435527",
|
| 340 |
+
"timestamp_ms": 1771243416435.539,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T21:03:41.442964",
|
| 346 |
+
"timestamp_ms": 1771243421442.9758,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T21:03:46.450602",
|
| 352 |
+
"timestamp_ms": 1771243426450.6143,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T21:03:51.459987",
|
| 358 |
+
"timestamp_ms": 1771243431460.0051,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T21:03:56.468497",
|
| 364 |
+
"timestamp_ms": 1771243436468.5117,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T21:04:01.476579",
|
| 370 |
+
"timestamp_ms": 1771243441476.591,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T21:04:06.484340",
|
| 376 |
+
"timestamp_ms": 1771243446484.3528,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T21:04:11.492288",
|
| 382 |
+
"timestamp_ms": 1771243451492.3032,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T21:04:16.499894",
|
| 388 |
+
"timestamp_ms": 1771243456499.906,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T21:04:21.508704",
|
| 394 |
+
"timestamp_ms": 1771243461508.7205,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T21:04:26.516989",
|
| 400 |
+
"timestamp_ms": 1771243466517.0024,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T21:04:31.524554",
|
| 406 |
+
"timestamp_ms": 1771243471524.5674,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T21:04:36.525232",
|
| 412 |
+
"timestamp_ms": 1771243476525.2688,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T21:04:36.525446",
|
| 418 |
+
"timestamp_ms": 1771243476525.4558,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3011,30.110466,1843.16453,61.214037,105.345,254.157,9.269
|
| 3 |
+
idle_post,3000,29.989964,1327.652118,44.258157,78.8503,78.96,8.114
|
| 4 |
+
idle_pre,3003,30.048817,1597.815237,53.201305,79.369,80.05,8.978
|
| 5 |
+
inference,3,0.029527,0.282394,9.564,9.564,9.564,9.564
|
| 6 |
+
inference_idle,30057,300.692095,23730.143473,78.917445,80.884,81.444,9.564
|
| 7 |
+
train,34,0.378608,3.486977,9.21,9.21,9.21,9.21
|
| 8 |
+
train_compute,2961,29.64147,9203.607656,310.741675,396.959,401.572,9.21
|
| 9 |
+
validation,209,2.166443,354.439076,166.603943,255.334,255.334,80.279
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_efficientnet_b0_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
| 2 |
+
3537,35437.483,train_compute,train_compute,135.564,278.162,142.598,9.97095,14301.345462,up
|
| 3 |
+
3587,35937.518,train_compute,train_compute,278.162,366.072,87.91,10.036018,8759.450231,up
|
| 4 |
+
5687,56937.517,train_compute,train_compute,401.572,290.622,-110.95,10.028068,-11063.945743,down
|
| 5 |
+
5737,57437.482,train_compute,train_compute,290.622,164.58,-126.042,9.963489,-12650.387804,down
|
| 6 |
+
5787,57937.481,train_compute,train_compute,164.58,98.759,-65.821,9.965377,-6604.968352,down
|
| 7 |
+
6076,60935.799,validation,validation,80.279,160.989,80.71,10.00119,8070.03966,up
|
| 8 |
+
6126,61435.796,validation,validation,160.989,255.334,94.345,9.999144,9435.307655,up
|
| 9 |
+
6276,62935.85,idle_mid,idle_mid,236.196,165.923,-70.273,10.043749,-6996.690178,down
|
| 10 |
+
6326,63435.8,idle_mid,idle_mid,165.923,105.345,-60.578,9.994563,-6061.095378,down
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_mobilenetv2_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_mobilenetv2_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "mobilenetv2",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_mobilenetv2_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:22:37.734623",
|
| 4 |
+
"timestamp_ms": 1771240957734.637,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:23:07.734795",
|
| 10 |
+
"timestamp_ms": 1771240987734.8955,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:23:07.762212",
|
| 16 |
+
"timestamp_ms": 1771240987762.221,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:23:21.397368",
|
| 22 |
+
"timestamp_ms": 1771241001397.399,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:23:22.123935",
|
| 28 |
+
"timestamp_ms": 1771241002123.945,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:23:22.124909",
|
| 34 |
+
"timestamp_ms": 1771241002124.9143,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:23:52.125066",
|
| 40 |
+
"timestamp_ms": 1771241032125.1084,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:23:52.157084",
|
| 46 |
+
"timestamp_ms": 1771241032157.106,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:23:52.170486",
|
| 52 |
+
"timestamp_ms": 1771241032170.4978,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:23:57.198906",
|
| 58 |
+
"timestamp_ms": 1771241037198.921,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:24:02.206012",
|
| 64 |
+
"timestamp_ms": 1771241042206.0276,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:24:07.212535",
|
| 70 |
+
"timestamp_ms": 1771241047212.5496,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:24:12.219096",
|
| 76 |
+
"timestamp_ms": 1771241052219.1108,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:24:17.225029",
|
| 82 |
+
"timestamp_ms": 1771241057225.0432,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:24:22.231484",
|
| 88 |
+
"timestamp_ms": 1771241062231.4988,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:24:27.236869",
|
| 94 |
+
"timestamp_ms": 1771241067236.884,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:24:32.243363",
|
| 100 |
+
"timestamp_ms": 1771241072243.3774,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:24:37.249029",
|
| 106 |
+
"timestamp_ms": 1771241077249.0405,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:24:42.255332",
|
| 112 |
+
"timestamp_ms": 1771241082255.3457,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:24:47.261218",
|
| 118 |
+
"timestamp_ms": 1771241087261.231,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:24:52.267808",
|
| 124 |
+
"timestamp_ms": 1771241092267.822,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:24:57.273829",
|
| 130 |
+
"timestamp_ms": 1771241097273.8523,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:25:02.279608",
|
| 136 |
+
"timestamp_ms": 1771241102279.6196,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:25:07.285681",
|
| 142 |
+
"timestamp_ms": 1771241107285.695,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:25:12.292521",
|
| 148 |
+
"timestamp_ms": 1771241112292.5361,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:25:17.298768",
|
| 154 |
+
"timestamp_ms": 1771241117298.7815,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:25:22.304646",
|
| 160 |
+
"timestamp_ms": 1771241122304.6597,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:25:27.310397",
|
| 166 |
+
"timestamp_ms": 1771241127310.4072,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:25:32.317264",
|
| 172 |
+
"timestamp_ms": 1771241132317.2769,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:25:37.323009",
|
| 178 |
+
"timestamp_ms": 1771241137323.023,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:25:42.328995",
|
| 184 |
+
"timestamp_ms": 1771241142329.0073,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:25:47.334956",
|
| 190 |
+
"timestamp_ms": 1771241147334.969,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:25:52.340658",
|
| 196 |
+
"timestamp_ms": 1771241152340.6682,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:25:57.346987",
|
| 202 |
+
"timestamp_ms": 1771241157346.9995,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:26:02.353164",
|
| 208 |
+
"timestamp_ms": 1771241162353.178,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:26:07.358965",
|
| 214 |
+
"timestamp_ms": 1771241167358.9785,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:26:12.364687",
|
| 220 |
+
"timestamp_ms": 1771241172364.7004,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:26:17.369713",
|
| 226 |
+
"timestamp_ms": 1771241177369.7244,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:26:22.375749",
|
| 232 |
+
"timestamp_ms": 1771241182375.762,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:26:27.381876",
|
| 238 |
+
"timestamp_ms": 1771241187381.8901,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:26:32.388135",
|
| 244 |
+
"timestamp_ms": 1771241192388.1484,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:26:37.394155",
|
| 250 |
+
"timestamp_ms": 1771241197394.1726,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:26:42.400009",
|
| 256 |
+
"timestamp_ms": 1771241202400.021,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:26:47.405971",
|
| 262 |
+
"timestamp_ms": 1771241207405.9868,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:26:52.412021",
|
| 268 |
+
"timestamp_ms": 1771241212412.0361,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:26:57.417672",
|
| 274 |
+
"timestamp_ms": 1771241217417.686,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:27:02.423618",
|
| 280 |
+
"timestamp_ms": 1771241222423.6323,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:27:07.429452",
|
| 286 |
+
"timestamp_ms": 1771241227429.4634,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:27:12.435745",
|
| 292 |
+
"timestamp_ms": 1771241232435.7588,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:27:17.443815",
|
| 298 |
+
"timestamp_ms": 1771241237443.8286,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:27:22.450655",
|
| 304 |
+
"timestamp_ms": 1771241242450.669,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:27:27.457033",
|
| 310 |
+
"timestamp_ms": 1771241247457.0466,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:27:32.463358",
|
| 316 |
+
"timestamp_ms": 1771241252463.3694,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:27:37.468763",
|
| 322 |
+
"timestamp_ms": 1771241257468.7773,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:27:42.474954",
|
| 328 |
+
"timestamp_ms": 1771241262474.9666,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:27:47.480260",
|
| 334 |
+
"timestamp_ms": 1771241267480.2734,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:27:52.486064",
|
| 340 |
+
"timestamp_ms": 1771241272486.0771,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:27:57.495598",
|
| 346 |
+
"timestamp_ms": 1771241277495.6113,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:28:02.501450",
|
| 352 |
+
"timestamp_ms": 1771241282501.4648,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:28:07.507915",
|
| 358 |
+
"timestamp_ms": 1771241287507.9287,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:28:12.514571",
|
| 364 |
+
"timestamp_ms": 1771241292514.5854,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:28:17.520513",
|
| 370 |
+
"timestamp_ms": 1771241297520.5269,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:28:22.526822",
|
| 376 |
+
"timestamp_ms": 1771241302526.8857,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:28:27.533646",
|
| 382 |
+
"timestamp_ms": 1771241307533.6602,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:28:32.539605",
|
| 388 |
+
"timestamp_ms": 1771241312539.619,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:28:37.545672",
|
| 394 |
+
"timestamp_ms": 1771241317545.6858,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:28:42.551135",
|
| 400 |
+
"timestamp_ms": 1771241322551.1462,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:28:47.557116",
|
| 406 |
+
"timestamp_ms": 1771241327557.1294,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:28:52.557690",
|
| 412 |
+
"timestamp_ms": 1771241332557.7234,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:28:52.557961",
|
| 418 |
+
"timestamp_ms": 1771241332557.9724,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet18",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T19:54:22.945900",
|
| 4 |
+
"timestamp_ms": 1771239262945.9136,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T19:54:52.946115",
|
| 10 |
+
"timestamp_ms": 1771239292946.1592,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T19:54:52.972347",
|
| 16 |
+
"timestamp_ms": 1771239292972.3564,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T19:55:02.341254",
|
| 22 |
+
"timestamp_ms": 1771239302341.2678,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T19:55:02.858168",
|
| 28 |
+
"timestamp_ms": 1771239302858.179,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T19:55:02.858661",
|
| 34 |
+
"timestamp_ms": 1771239302858.6655,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T19:55:32.858807",
|
| 40 |
+
"timestamp_ms": 1771239332858.8938,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T19:55:32.884857",
|
| 46 |
+
"timestamp_ms": 1771239332884.8813,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T19:55:32.896630",
|
| 52 |
+
"timestamp_ms": 1771239332896.6418,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T19:55:37.918843",
|
| 58 |
+
"timestamp_ms": 1771239337918.8584,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T19:55:42.923643",
|
| 64 |
+
"timestamp_ms": 1771239342923.6567,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T19:55:47.927378",
|
| 70 |
+
"timestamp_ms": 1771239347927.3909,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T19:55:52.932008",
|
| 76 |
+
"timestamp_ms": 1771239352932.0266,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T19:55:57.936088",
|
| 82 |
+
"timestamp_ms": 1771239357936.1042,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T19:56:02.940856",
|
| 88 |
+
"timestamp_ms": 1771239362940.8699,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T19:56:07.945057",
|
| 94 |
+
"timestamp_ms": 1771239367945.0718,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T19:56:12.950024",
|
| 100 |
+
"timestamp_ms": 1771239372950.0447,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T19:56:17.955187",
|
| 106 |
+
"timestamp_ms": 1771239377955.204,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T19:56:22.959625",
|
| 112 |
+
"timestamp_ms": 1771239382959.6438,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T19:56:27.964509",
|
| 118 |
+
"timestamp_ms": 1771239387964.524,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T19:56:32.969167",
|
| 124 |
+
"timestamp_ms": 1771239392969.1868,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T19:56:37.974086",
|
| 130 |
+
"timestamp_ms": 1771239397974.1082,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T19:56:42.978521",
|
| 136 |
+
"timestamp_ms": 1771239402978.5347,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T19:56:47.983716",
|
| 142 |
+
"timestamp_ms": 1771239407983.7295,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T19:56:52.988325",
|
| 148 |
+
"timestamp_ms": 1771239412988.339,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T19:56:57.993296",
|
| 154 |
+
"timestamp_ms": 1771239417993.3098,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T19:57:02.998267",
|
| 160 |
+
"timestamp_ms": 1771239422998.279,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T19:57:08.002844",
|
| 166 |
+
"timestamp_ms": 1771239428002.858,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T19:57:13.008258",
|
| 172 |
+
"timestamp_ms": 1771239433008.2722,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T19:57:18.013011",
|
| 178 |
+
"timestamp_ms": 1771239438013.031,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T19:57:23.017087",
|
| 184 |
+
"timestamp_ms": 1771239443017.1013,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T19:57:28.021857",
|
| 190 |
+
"timestamp_ms": 1771239448021.871,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T19:57:33.026432",
|
| 196 |
+
"timestamp_ms": 1771239453026.4524,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T19:57:38.030973",
|
| 202 |
+
"timestamp_ms": 1771239458030.9937,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T19:57:43.034989",
|
| 208 |
+
"timestamp_ms": 1771239463035.0051,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T19:57:48.038822",
|
| 214 |
+
"timestamp_ms": 1771239468038.85,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T19:57:53.043255",
|
| 220 |
+
"timestamp_ms": 1771239473043.2695,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T19:57:58.047917",
|
| 226 |
+
"timestamp_ms": 1771239478047.9316,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T19:58:03.053603",
|
| 232 |
+
"timestamp_ms": 1771239483053.6177,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T19:58:08.058941",
|
| 238 |
+
"timestamp_ms": 1771239488058.9556,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T19:58:13.062467",
|
| 244 |
+
"timestamp_ms": 1771239493062.4844,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T19:58:18.066640",
|
| 250 |
+
"timestamp_ms": 1771239498066.654,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T19:58:23.070597",
|
| 256 |
+
"timestamp_ms": 1771239503070.6135,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T19:58:28.075581",
|
| 262 |
+
"timestamp_ms": 1771239508075.5945,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T19:58:33.079790",
|
| 268 |
+
"timestamp_ms": 1771239513079.8057,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T19:58:38.084416",
|
| 274 |
+
"timestamp_ms": 1771239518084.428,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T19:58:43.089625",
|
| 280 |
+
"timestamp_ms": 1771239523089.6392,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T19:58:48.094657",
|
| 286 |
+
"timestamp_ms": 1771239528094.6711,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T19:58:53.099406",
|
| 292 |
+
"timestamp_ms": 1771239533099.4268,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T19:58:58.104379",
|
| 298 |
+
"timestamp_ms": 1771239538104.4045,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T19:59:03.109209",
|
| 304 |
+
"timestamp_ms": 1771239543109.223,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T19:59:08.113796",
|
| 310 |
+
"timestamp_ms": 1771239548113.814,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T19:59:13.118123",
|
| 316 |
+
"timestamp_ms": 1771239553118.1375,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T19:59:18.123718",
|
| 322 |
+
"timestamp_ms": 1771239558123.7395,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T19:59:23.127806",
|
| 328 |
+
"timestamp_ms": 1771239563127.8198,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T19:59:28.131867",
|
| 334 |
+
"timestamp_ms": 1771239568131.8833,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T19:59:33.136352",
|
| 340 |
+
"timestamp_ms": 1771239573136.3667,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T19:59:38.140718",
|
| 346 |
+
"timestamp_ms": 1771239578140.7356,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T19:59:43.144841",
|
| 352 |
+
"timestamp_ms": 1771239583144.8596,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T19:59:48.149338",
|
| 358 |
+
"timestamp_ms": 1771239588149.3552,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T19:59:53.154105",
|
| 364 |
+
"timestamp_ms": 1771239593154.119,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T19:59:58.159397",
|
| 370 |
+
"timestamp_ms": 1771239598159.4111,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:00:03.164191",
|
| 376 |
+
"timestamp_ms": 1771239603164.2068,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:00:08.168247",
|
| 382 |
+
"timestamp_ms": 1771239608168.2646,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:00:13.173230",
|
| 388 |
+
"timestamp_ms": 1771239613173.25,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:00:18.177795",
|
| 394 |
+
"timestamp_ms": 1771239618177.808,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:00:23.182540",
|
| 400 |
+
"timestamp_ms": 1771239623182.5544,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:00:28.187629",
|
| 406 |
+
"timestamp_ms": 1771239628187.6428,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:00:33.188282",
|
| 412 |
+
"timestamp_ms": 1771239633188.3152,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:00:33.188486",
|
| 418 |
+
"timestamp_ms": 1771239633188.4956,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3002,30.020032,1633.251173,54.405513,81.293,94.486,8.152
|
| 3 |
+
idle_post,3000,29.989965,1265.22156,42.177128,78.13225,78.431,8.059
|
| 4 |
+
idle_pre,3003,30.049034,1587.709157,52.865143,78.504,78.578,7.606
|
| 5 |
+
inference,1,0.009922,0.08408,8.474,8.474,8.474,8.474
|
| 6 |
+
inference_idle,30018,300.294939,23399.480477,77.921632,78.777,79.083,8.474
|
| 7 |
+
train,6,0.102379,0.888959,8.683,8.683,8.683,8.683
|
| 8 |
+
train_compute,925,9.27713,944.579334,101.885756,147.429,147.429,8.683
|
| 9 |
+
validation,41,0.493143,39.634463,80.653854,82.592,82.592,78.98
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet18_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnet50",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:01:08.765820",
|
| 4 |
+
"timestamp_ms": 1771239668765.8406,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:01:38.765981",
|
| 10 |
+
"timestamp_ms": 1771239698766.0251,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:01:38.808634",
|
| 16 |
+
"timestamp_ms": 1771239698808.6426,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:01:50.913639",
|
| 22 |
+
"timestamp_ms": 1771239710913.6704,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:01:51.730735",
|
| 28 |
+
"timestamp_ms": 1771239711730.7444,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:01:51.731556",
|
| 34 |
+
"timestamp_ms": 1771239711731.561,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:02:21.731704",
|
| 40 |
+
"timestamp_ms": 1771239741731.7466,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:02:21.758557",
|
| 46 |
+
"timestamp_ms": 1771239741758.575,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:02:21.773763",
|
| 52 |
+
"timestamp_ms": 1771239741773.775,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:02:26.802410",
|
| 58 |
+
"timestamp_ms": 1771239746802.425,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:02:31.809155",
|
| 64 |
+
"timestamp_ms": 1771239751809.1675,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:02:36.815700",
|
| 70 |
+
"timestamp_ms": 1771239756815.714,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:02:41.824352",
|
| 76 |
+
"timestamp_ms": 1771239761824.3652,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:02:46.830672",
|
| 82 |
+
"timestamp_ms": 1771239766830.6843,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:02:51.837571",
|
| 88 |
+
"timestamp_ms": 1771239771837.5833,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:02:56.844129",
|
| 94 |
+
"timestamp_ms": 1771239776844.141,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:03:01.850068",
|
| 100 |
+
"timestamp_ms": 1771239781850.0823,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:03:06.856626",
|
| 106 |
+
"timestamp_ms": 1771239786856.6384,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:03:11.863187",
|
| 112 |
+
"timestamp_ms": 1771239791863.2007,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:03:16.870503",
|
| 118 |
+
"timestamp_ms": 1771239796870.5168,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:03:21.877215",
|
| 124 |
+
"timestamp_ms": 1771239801877.23,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:03:26.884242",
|
| 130 |
+
"timestamp_ms": 1771239806884.2556,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:03:31.890539",
|
| 136 |
+
"timestamp_ms": 1771239811890.5525,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:03:36.897709",
|
| 142 |
+
"timestamp_ms": 1771239816897.7224,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:03:41.901984",
|
| 148 |
+
"timestamp_ms": 1771239821901.9946,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:03:46.909414",
|
| 154 |
+
"timestamp_ms": 1771239826909.429,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:03:51.916020",
|
| 160 |
+
"timestamp_ms": 1771239831916.0342,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:03:56.922699",
|
| 166 |
+
"timestamp_ms": 1771239836922.713,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:04:01.929001",
|
| 172 |
+
"timestamp_ms": 1771239841929.0164,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:04:06.935082",
|
| 178 |
+
"timestamp_ms": 1771239846935.0928,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:04:11.944718",
|
| 184 |
+
"timestamp_ms": 1771239851944.7327,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:04:16.952208",
|
| 190 |
+
"timestamp_ms": 1771239856952.2224,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:04:21.958921",
|
| 196 |
+
"timestamp_ms": 1771239861958.9338,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:04:26.965945",
|
| 202 |
+
"timestamp_ms": 1771239866965.9578,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:04:31.971742",
|
| 208 |
+
"timestamp_ms": 1771239871971.7546,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:04:36.979161",
|
| 214 |
+
"timestamp_ms": 1771239876979.178,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:04:41.985824",
|
| 220 |
+
"timestamp_ms": 1771239881985.8438,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:04:46.992751",
|
| 226 |
+
"timestamp_ms": 1771239886992.764,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:04:52.001487",
|
| 232 |
+
"timestamp_ms": 1771239892001.5005,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:04:57.007443",
|
| 238 |
+
"timestamp_ms": 1771239897007.4563,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:05:02.013685",
|
| 244 |
+
"timestamp_ms": 1771239902013.699,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:05:07.020111",
|
| 250 |
+
"timestamp_ms": 1771239907020.1238,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:05:12.026786",
|
| 256 |
+
"timestamp_ms": 1771239912026.7996,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:05:17.032946",
|
| 262 |
+
"timestamp_ms": 1771239917032.958,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:05:22.042794",
|
| 268 |
+
"timestamp_ms": 1771239922042.8071,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:05:27.049235",
|
| 274 |
+
"timestamp_ms": 1771239927049.248,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:05:32.057986",
|
| 280 |
+
"timestamp_ms": 1771239932057.999,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:05:37.063933",
|
| 286 |
+
"timestamp_ms": 1771239937063.9465,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:05:42.070063",
|
| 292 |
+
"timestamp_ms": 1771239942070.0735,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:05:47.078062",
|
| 298 |
+
"timestamp_ms": 1771239947078.078,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:05:52.084771",
|
| 304 |
+
"timestamp_ms": 1771239952084.7842,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:05:57.091556",
|
| 310 |
+
"timestamp_ms": 1771239957091.5693,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:06:02.098111",
|
| 316 |
+
"timestamp_ms": 1771239962098.1255,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:06:07.104820",
|
| 322 |
+
"timestamp_ms": 1771239967104.8306,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:06:12.114315",
|
| 328 |
+
"timestamp_ms": 1771239972114.3284,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:06:17.120435",
|
| 334 |
+
"timestamp_ms": 1771239977120.4487,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:06:22.127793",
|
| 340 |
+
"timestamp_ms": 1771239982127.8093,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:06:27.134125",
|
| 346 |
+
"timestamp_ms": 1771239987134.1375,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:06:32.140961",
|
| 352 |
+
"timestamp_ms": 1771239992140.975,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:06:37.146877",
|
| 358 |
+
"timestamp_ms": 1771239997146.8901,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:06:42.153670",
|
| 364 |
+
"timestamp_ms": 1771240002153.6895,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:06:47.160133",
|
| 370 |
+
"timestamp_ms": 1771240007160.147,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:06:52.168045",
|
| 376 |
+
"timestamp_ms": 1771240012168.0552,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:06:57.174633",
|
| 382 |
+
"timestamp_ms": 1771240017174.6443,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:07:02.182067",
|
| 388 |
+
"timestamp_ms": 1771240022182.0808,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:07:07.188768",
|
| 394 |
+
"timestamp_ms": 1771240027188.7812,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:07:12.195240",
|
| 400 |
+
"timestamp_ms": 1771240032195.2527,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:07:17.201263",
|
| 406 |
+
"timestamp_ms": 1771240037201.2742,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:07:22.201915",
|
| 412 |
+
"timestamp_ms": 1771240042201.948,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:07:22.202113",
|
| 418 |
+
"timestamp_ms": 1771240042202.1223,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3002,30.019992,1672.888168,55.725697,87.069,121.434,7.85
|
| 3 |
+
idle_post,3000,29.990041,1238.822888,41.296948,78.54785,78.65,7.744
|
| 4 |
+
idle_pre,3005,30.083658,1615.429094,53.747655,78.881,78.924,8.469
|
| 5 |
+
inference,2,0.020002,0.159373,7.968,7.968,7.968,7.968
|
| 6 |
+
inference_idle,30033,300.432184,23438.569513,78.01625,79.001,79.576,7.968
|
| 7 |
+
train,5,0.082202,0.748449,9.105,9.105,9.105,9.105
|
| 8 |
+
train_compute,1199,12.018986,1617.81513,134.698481,173.34,173.449,9.105
|
| 9 |
+
validation,71,0.788471,73.145081,92.396831,103.7905,111.454,89.604
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnet50_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "resnext50_32x4d",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:07:57.842447",
|
| 4 |
+
"timestamp_ms": 1771240077842.4592,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:08:27.842666",
|
| 10 |
+
"timestamp_ms": 1771240107842.7092,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:08:27.875706",
|
| 16 |
+
"timestamp_ms": 1771240107875.7148,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:09:10.957441",
|
| 22 |
+
"timestamp_ms": 1771240150957.4587,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:09:13.469463",
|
| 28 |
+
"timestamp_ms": 1771240153469.477,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:09:13.470649",
|
| 34 |
+
"timestamp_ms": 1771240153470.6597,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:09:43.470887",
|
| 40 |
+
"timestamp_ms": 1771240183470.9297,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:09:43.632924",
|
| 46 |
+
"timestamp_ms": 1771240183632.9626,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:09:43.661985",
|
| 52 |
+
"timestamp_ms": 1771240183661.9978,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:09:48.851983",
|
| 58 |
+
"timestamp_ms": 1771240188852.002,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:09:53.863433",
|
| 64 |
+
"timestamp_ms": 1771240193863.4453,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:09:58.875318",
|
| 70 |
+
"timestamp_ms": 1771240198875.3303,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:10:03.887335",
|
| 76 |
+
"timestamp_ms": 1771240203887.3508,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:10:08.899475",
|
| 82 |
+
"timestamp_ms": 1771240208899.4873,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:10:13.911391",
|
| 88 |
+
"timestamp_ms": 1771240213911.4033,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:10:18.922998",
|
| 94 |
+
"timestamp_ms": 1771240218923.0093,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:10:23.934875",
|
| 100 |
+
"timestamp_ms": 1771240223934.8884,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:10:28.946674",
|
| 106 |
+
"timestamp_ms": 1771240228946.6873,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:10:33.958758",
|
| 112 |
+
"timestamp_ms": 1771240233958.7712,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:10:38.970451",
|
| 118 |
+
"timestamp_ms": 1771240238970.4631,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:10:43.982031",
|
| 124 |
+
"timestamp_ms": 1771240243982.0435,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:10:48.994022",
|
| 130 |
+
"timestamp_ms": 1771240248994.0337,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:10:54.005991",
|
| 136 |
+
"timestamp_ms": 1771240254006.004,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:10:59.017995",
|
| 142 |
+
"timestamp_ms": 1771240259018.0066,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:11:04.030121",
|
| 148 |
+
"timestamp_ms": 1771240264030.1338,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:11:09.041815",
|
| 154 |
+
"timestamp_ms": 1771240269041.829,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:11:14.053763",
|
| 160 |
+
"timestamp_ms": 1771240274053.7751,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:11:19.065674",
|
| 166 |
+
"timestamp_ms": 1771240279065.686,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:11:24.077467",
|
| 172 |
+
"timestamp_ms": 1771240284077.479,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:11:29.089591",
|
| 178 |
+
"timestamp_ms": 1771240289089.6072,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:11:34.101314",
|
| 184 |
+
"timestamp_ms": 1771240294101.3252,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:11:39.112970",
|
| 190 |
+
"timestamp_ms": 1771240299112.9834,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:11:44.124900",
|
| 196 |
+
"timestamp_ms": 1771240304124.9146,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:11:49.137033",
|
| 202 |
+
"timestamp_ms": 1771240309137.0479,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:11:54.148815",
|
| 208 |
+
"timestamp_ms": 1771240314148.8271,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:11:59.161148",
|
| 214 |
+
"timestamp_ms": 1771240319161.1606,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:12:04.173066",
|
| 220 |
+
"timestamp_ms": 1771240324173.0784,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:12:09.185110",
|
| 226 |
+
"timestamp_ms": 1771240329185.1226,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:12:14.196930",
|
| 232 |
+
"timestamp_ms": 1771240334196.942,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:12:19.208844",
|
| 238 |
+
"timestamp_ms": 1771240339208.8555,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:12:24.220682",
|
| 244 |
+
"timestamp_ms": 1771240344220.695,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:12:29.232695",
|
| 250 |
+
"timestamp_ms": 1771240349232.7063,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:12:34.244387",
|
| 256 |
+
"timestamp_ms": 1771240354244.4011,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:12:39.257283",
|
| 262 |
+
"timestamp_ms": 1771240359257.307,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:12:44.269445",
|
| 268 |
+
"timestamp_ms": 1771240364269.4568,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:12:49.281105",
|
| 274 |
+
"timestamp_ms": 1771240369281.1174,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:12:54.293061",
|
| 280 |
+
"timestamp_ms": 1771240374293.073,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:12:59.304990",
|
| 286 |
+
"timestamp_ms": 1771240379305.0017,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:13:04.317032",
|
| 292 |
+
"timestamp_ms": 1771240384317.044,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:13:09.328811",
|
| 298 |
+
"timestamp_ms": 1771240389328.8242,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:13:14.340676",
|
| 304 |
+
"timestamp_ms": 1771240394340.6887,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:13:19.352611",
|
| 310 |
+
"timestamp_ms": 1771240399352.622,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:13:24.364483",
|
| 316 |
+
"timestamp_ms": 1771240404364.4949,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:13:29.375903",
|
| 322 |
+
"timestamp_ms": 1771240409375.9177,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:13:34.387674",
|
| 328 |
+
"timestamp_ms": 1771240414387.6853,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:13:39.399478",
|
| 334 |
+
"timestamp_ms": 1771240419399.493,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:13:44.411319",
|
| 340 |
+
"timestamp_ms": 1771240424411.334,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:13:49.424193",
|
| 346 |
+
"timestamp_ms": 1771240429424.2122,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:13:54.436106",
|
| 352 |
+
"timestamp_ms": 1771240434436.1182,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:13:59.447903",
|
| 358 |
+
"timestamp_ms": 1771240439447.914,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:14:04.459686",
|
| 364 |
+
"timestamp_ms": 1771240444459.6985,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:14:09.471391",
|
| 370 |
+
"timestamp_ms": 1771240449471.4023,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:14:14.483394",
|
| 376 |
+
"timestamp_ms": 1771240454483.4065,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:14:19.495299",
|
| 382 |
+
"timestamp_ms": 1771240459495.312,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:14:24.507102",
|
| 388 |
+
"timestamp_ms": 1771240464507.1172,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:14:29.518751",
|
| 394 |
+
"timestamp_ms": 1771240469518.7634,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:14:34.530628",
|
| 400 |
+
"timestamp_ms": 1771240474530.644,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:14:39.542567",
|
| 406 |
+
"timestamp_ms": 1771240479542.5837,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:14:44.543237",
|
| 412 |
+
"timestamp_ms": 1771240484543.2742,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:14:44.543454",
|
| 418 |
+
"timestamp_ms": 1771240484543.4631,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/gpu_samples.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/phase_power_summary.csv
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
phase,num_samples,duration_s,energy_j,power_avg_w,power_p95_w,power_peak_w,power_min_w
|
| 2 |
+
idle_mid,3015,30.153913,2067.205693,68.56171,162.888,449.579,8.323
|
| 3 |
+
idle_post,3000,29.990032,1307.793388,43.596434,78.4433,78.718,8.621
|
| 4 |
+
idle_pre,3004,30.072836,1603.967374,53.384863,78.634,79.244,7.672
|
| 5 |
+
inference,4,0.039042,0.576684,14.771,14.771,14.771,14.771
|
| 6 |
+
inference_idle,30077,300.880658,23924.409513,79.514501,83.55,84.067,14.771
|
| 7 |
+
train,32,0.35083,3.923633,11.431406,26.721,26.721,8.6
|
| 8 |
+
train_compute,4270,42.732029,16277.21075,381.140643,512.471,514.289,26.721
|
| 9 |
+
validation,240,2.48003,695.916299,287.397533,449.579,449.579,83.336
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_resnext50_32x4d_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/samples/transition_points.csv
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
idx,timestamp_ms,phase_from,phase_to,power_from_w,power_to_w,delta_power_w,delta_time_ms,dpdt_w_per_s,direction
|
| 2 |
+
3130,31374.261,train_compute,train_compute,44.551,135.18,90.629,10.085456,8986.108248,up
|
| 3 |
+
3830,38375.069,train_compute,train_compute,130.066,300.216,170.15,10.860761,15666.489771,up
|
| 4 |
+
3880,38874.213,train_compute,train_compute,300.216,508.229,208.013,10.036647,20725.347879,up
|
| 5 |
+
6730,67374.182,train_compute,train_compute,512.197,449.993,-62.204,9.973721,-6236.789662,down
|
| 6 |
+
6780,67874.19,train_compute,train_compute,449.993,259.411,-190.582,9.982005,-19092.557067,down
|
| 7 |
+
6830,68374.207,train_compute,train_compute,259.411,96.888,-162.523,10.033342,-16198.291566,down
|
| 8 |
+
7369,73876.314,validation,validation,84.47,219.647,135.177,10.051331,13448.666626,up
|
| 9 |
+
7419,74379.556,validation,validation,219.647,371.249,151.602,13.285373,11411.196369,up
|
| 10 |
+
7469,74876.259,validation,validation,371.249,447.027,75.778,8.801154,8610.007301,up
|
| 11 |
+
7569,75876.298,idle_mid,idle_mid,449.579,388.361,-61.218,10.033403,-6101.419394,down
|
| 12 |
+
7619,76376.273,idle_mid,idle_mid,388.361,205.155,-183.206,9.975564,-18365.477727,down
|
| 13 |
+
7769,77876.314,idle_mid,idle_mid,162.888,91.937,-70.951,10.054086,-7056.931888,down
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"sampling_interval_ms": 10.0,
|
| 3 |
+
"idle_pre_duration": 30.0,
|
| 4 |
+
"idle_mid_duration": 30.0,
|
| 5 |
+
"idle_post_duration": 30.0,
|
| 6 |
+
"train_epochs": 1,
|
| 7 |
+
"batch_size": 64,
|
| 8 |
+
"learning_rate": 0.01,
|
| 9 |
+
"num_workers": 8,
|
| 10 |
+
"use_amp": true,
|
| 11 |
+
"train_time_limit_sec": null,
|
| 12 |
+
"target_val_acc_pct": null,
|
| 13 |
+
"checkpoint_every_steps": null,
|
| 14 |
+
"finegrained_train_phases": true,
|
| 15 |
+
"automl_n_trials": 10,
|
| 16 |
+
"automl_max_epochs_per_trial": 3,
|
| 17 |
+
"automl_pruning": true,
|
| 18 |
+
"diffusion_train_epochs": 5,
|
| 19 |
+
"diffusion_train_steps_per_epoch": 20,
|
| 20 |
+
"diffusion_train_batch_size": 1,
|
| 21 |
+
"diffusion_train_lr": 0.0001,
|
| 22 |
+
"diffusion_steps": 30,
|
| 23 |
+
"inference_total_requests": 10000,
|
| 24 |
+
"inference_interval_sec": 5.0,
|
| 25 |
+
"inference_pattern": "fixed",
|
| 26 |
+
"inference_variable_min": 2.0,
|
| 27 |
+
"inference_variable_max": 10.0,
|
| 28 |
+
"inference_burst_count": 5,
|
| 29 |
+
"inference_burst_gap": 1.0,
|
| 30 |
+
"inference_burst_pause": 15.0,
|
| 31 |
+
"inference_duration_sec": 300.0,
|
| 32 |
+
"clock_lock_gpu": null,
|
| 33 |
+
"clock_lock_mem": null,
|
| 34 |
+
"ramp_enabled": false,
|
| 35 |
+
"ramp_start_w": null,
|
| 36 |
+
"ramp_end_w": null,
|
| 37 |
+
"ramp_steps": 5,
|
| 38 |
+
"ramp_step_duration": 3.0,
|
| 39 |
+
"cuda_index": 0,
|
| 40 |
+
"gpu_physical_index": 0,
|
| 41 |
+
"device": "cuda:0",
|
| 42 |
+
"seed": 42,
|
| 43 |
+
"dataset_allow_download": false,
|
| 44 |
+
"dataset_root": "/data/datasets",
|
| 45 |
+
"imagenet_root": "/data/datasets/imagenet",
|
| 46 |
+
"transition_dpdt_threshold_w_per_s": 5000.0,
|
| 47 |
+
"checkpoint_dir": "/home/hyunwoo/DataCenter_train/hyp_opt/outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/checkpoints",
|
| 48 |
+
"mode": "fixed",
|
| 49 |
+
"model": "swin_t",
|
| 50 |
+
"dataset": "cifar10",
|
| 51 |
+
"power_cap": null
|
| 52 |
+
}
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/phase_transitions.json
ADDED
|
@@ -0,0 +1,422 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"timestamp": "2026-02-16T20:50:52.816273",
|
| 4 |
+
"timestamp_ms": 1771242652816.2935,
|
| 5 |
+
"phase": "idle_pre",
|
| 6 |
+
"sub_phase": ""
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"timestamp": "2026-02-16T20:51:22.816453",
|
| 10 |
+
"timestamp_ms": 1771242682816.495,
|
| 11 |
+
"phase": "train",
|
| 12 |
+
"sub_phase": "start"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"timestamp": "2026-02-16T20:51:22.861040",
|
| 16 |
+
"timestamp_ms": 1771242682861.049,
|
| 17 |
+
"phase": "train",
|
| 18 |
+
"sub_phase": "epoch_1_start"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"timestamp": "2026-02-16T20:51:54.078341",
|
| 22 |
+
"timestamp_ms": 1771242714078.3545,
|
| 23 |
+
"phase": "validation",
|
| 24 |
+
"sub_phase": "epoch_1"
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"timestamp": "2026-02-16T20:51:56.604321",
|
| 28 |
+
"timestamp_ms": 1771242716604.331,
|
| 29 |
+
"phase": "train",
|
| 30 |
+
"sub_phase": "finished"
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"timestamp": "2026-02-16T20:51:56.605428",
|
| 34 |
+
"timestamp_ms": 1771242716605.434,
|
| 35 |
+
"phase": "idle_mid",
|
| 36 |
+
"sub_phase": ""
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"timestamp": "2026-02-16T20:52:26.605629",
|
| 40 |
+
"timestamp_ms": 1771242746605.6716,
|
| 41 |
+
"phase": "inference",
|
| 42 |
+
"sub_phase": "start"
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"timestamp": "2026-02-16T20:52:26.796575",
|
| 46 |
+
"timestamp_ms": 1771242746796.6086,
|
| 47 |
+
"phase": "inference",
|
| 48 |
+
"sub_phase": "start_pattern_fixed"
|
| 49 |
+
},
|
| 50 |
+
{
|
| 51 |
+
"timestamp": "2026-02-16T20:52:26.820174",
|
| 52 |
+
"timestamp_ms": 1771242746820.1836,
|
| 53 |
+
"phase": "inference_idle",
|
| 54 |
+
"sub_phase": "req_0"
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"timestamp": "2026-02-16T20:52:32.045753",
|
| 58 |
+
"timestamp_ms": 1771242752045.7698,
|
| 59 |
+
"phase": "inference_idle",
|
| 60 |
+
"sub_phase": "req_1"
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"timestamp": "2026-02-16T20:52:37.059402",
|
| 64 |
+
"timestamp_ms": 1771242757059.415,
|
| 65 |
+
"phase": "inference_idle",
|
| 66 |
+
"sub_phase": "req_2"
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"timestamp": "2026-02-16T20:52:42.072844",
|
| 70 |
+
"timestamp_ms": 1771242762072.8608,
|
| 71 |
+
"phase": "inference_idle",
|
| 72 |
+
"sub_phase": "req_3"
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"timestamp": "2026-02-16T20:52:47.086429",
|
| 76 |
+
"timestamp_ms": 1771242767086.442,
|
| 77 |
+
"phase": "inference_idle",
|
| 78 |
+
"sub_phase": "req_4"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"timestamp": "2026-02-16T20:52:52.099937",
|
| 82 |
+
"timestamp_ms": 1771242772099.9543,
|
| 83 |
+
"phase": "inference_idle",
|
| 84 |
+
"sub_phase": "req_5"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"timestamp": "2026-02-16T20:52:57.113438",
|
| 88 |
+
"timestamp_ms": 1771242777113.4512,
|
| 89 |
+
"phase": "inference_idle",
|
| 90 |
+
"sub_phase": "req_6"
|
| 91 |
+
},
|
| 92 |
+
{
|
| 93 |
+
"timestamp": "2026-02-16T20:53:02.126850",
|
| 94 |
+
"timestamp_ms": 1771242782126.8625,
|
| 95 |
+
"phase": "inference_idle",
|
| 96 |
+
"sub_phase": "req_7"
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"timestamp": "2026-02-16T20:53:07.140380",
|
| 100 |
+
"timestamp_ms": 1771242787140.396,
|
| 101 |
+
"phase": "inference_idle",
|
| 102 |
+
"sub_phase": "req_8"
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"timestamp": "2026-02-16T20:53:12.154116",
|
| 106 |
+
"timestamp_ms": 1771242792154.1284,
|
| 107 |
+
"phase": "inference_idle",
|
| 108 |
+
"sub_phase": "req_9"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"timestamp": "2026-02-16T20:53:17.167886",
|
| 112 |
+
"timestamp_ms": 1771242797167.9016,
|
| 113 |
+
"phase": "inference_idle",
|
| 114 |
+
"sub_phase": "req_10"
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"timestamp": "2026-02-16T20:53:22.181251",
|
| 118 |
+
"timestamp_ms": 1771242802181.264,
|
| 119 |
+
"phase": "inference_idle",
|
| 120 |
+
"sub_phase": "req_11"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"timestamp": "2026-02-16T20:53:27.194376",
|
| 124 |
+
"timestamp_ms": 1771242807194.391,
|
| 125 |
+
"phase": "inference_idle",
|
| 126 |
+
"sub_phase": "req_12"
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"timestamp": "2026-02-16T20:53:32.207968",
|
| 130 |
+
"timestamp_ms": 1771242812207.9817,
|
| 131 |
+
"phase": "inference_idle",
|
| 132 |
+
"sub_phase": "req_13"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"timestamp": "2026-02-16T20:53:37.221250",
|
| 136 |
+
"timestamp_ms": 1771242817221.267,
|
| 137 |
+
"phase": "inference_idle",
|
| 138 |
+
"sub_phase": "req_14"
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"timestamp": "2026-02-16T20:53:42.234991",
|
| 142 |
+
"timestamp_ms": 1771242822235.0051,
|
| 143 |
+
"phase": "inference_idle",
|
| 144 |
+
"sub_phase": "req_15"
|
| 145 |
+
},
|
| 146 |
+
{
|
| 147 |
+
"timestamp": "2026-02-16T20:53:47.248717",
|
| 148 |
+
"timestamp_ms": 1771242827248.7302,
|
| 149 |
+
"phase": "inference_idle",
|
| 150 |
+
"sub_phase": "req_16"
|
| 151 |
+
},
|
| 152 |
+
{
|
| 153 |
+
"timestamp": "2026-02-16T20:53:52.262005",
|
| 154 |
+
"timestamp_ms": 1771242832262.017,
|
| 155 |
+
"phase": "inference_idle",
|
| 156 |
+
"sub_phase": "req_17"
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"timestamp": "2026-02-16T20:53:57.275408",
|
| 160 |
+
"timestamp_ms": 1771242837275.4207,
|
| 161 |
+
"phase": "inference_idle",
|
| 162 |
+
"sub_phase": "req_18"
|
| 163 |
+
},
|
| 164 |
+
{
|
| 165 |
+
"timestamp": "2026-02-16T20:54:02.289311",
|
| 166 |
+
"timestamp_ms": 1771242842289.3267,
|
| 167 |
+
"phase": "inference_idle",
|
| 168 |
+
"sub_phase": "req_19"
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"timestamp": "2026-02-16T20:54:07.302734",
|
| 172 |
+
"timestamp_ms": 1771242847302.7463,
|
| 173 |
+
"phase": "inference_idle",
|
| 174 |
+
"sub_phase": "req_20"
|
| 175 |
+
},
|
| 176 |
+
{
|
| 177 |
+
"timestamp": "2026-02-16T20:54:12.317203",
|
| 178 |
+
"timestamp_ms": 1771242852317.2153,
|
| 179 |
+
"phase": "inference_idle",
|
| 180 |
+
"sub_phase": "req_21"
|
| 181 |
+
},
|
| 182 |
+
{
|
| 183 |
+
"timestamp": "2026-02-16T20:54:17.330753",
|
| 184 |
+
"timestamp_ms": 1771242857330.7656,
|
| 185 |
+
"phase": "inference_idle",
|
| 186 |
+
"sub_phase": "req_22"
|
| 187 |
+
},
|
| 188 |
+
{
|
| 189 |
+
"timestamp": "2026-02-16T20:54:22.344349",
|
| 190 |
+
"timestamp_ms": 1771242862344.3647,
|
| 191 |
+
"phase": "inference_idle",
|
| 192 |
+
"sub_phase": "req_23"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"timestamp": "2026-02-16T20:54:27.357834",
|
| 196 |
+
"timestamp_ms": 1771242867357.8464,
|
| 197 |
+
"phase": "inference_idle",
|
| 198 |
+
"sub_phase": "req_24"
|
| 199 |
+
},
|
| 200 |
+
{
|
| 201 |
+
"timestamp": "2026-02-16T20:54:32.371288",
|
| 202 |
+
"timestamp_ms": 1771242872371.3027,
|
| 203 |
+
"phase": "inference_idle",
|
| 204 |
+
"sub_phase": "req_25"
|
| 205 |
+
},
|
| 206 |
+
{
|
| 207 |
+
"timestamp": "2026-02-16T20:54:37.384958",
|
| 208 |
+
"timestamp_ms": 1771242877384.971,
|
| 209 |
+
"phase": "inference_idle",
|
| 210 |
+
"sub_phase": "req_26"
|
| 211 |
+
},
|
| 212 |
+
{
|
| 213 |
+
"timestamp": "2026-02-16T20:54:42.398790",
|
| 214 |
+
"timestamp_ms": 1771242882398.803,
|
| 215 |
+
"phase": "inference_idle",
|
| 216 |
+
"sub_phase": "req_27"
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"timestamp": "2026-02-16T20:54:47.412409",
|
| 220 |
+
"timestamp_ms": 1771242887412.4216,
|
| 221 |
+
"phase": "inference_idle",
|
| 222 |
+
"sub_phase": "req_28"
|
| 223 |
+
},
|
| 224 |
+
{
|
| 225 |
+
"timestamp": "2026-02-16T20:54:52.425745",
|
| 226 |
+
"timestamp_ms": 1771242892425.7603,
|
| 227 |
+
"phase": "inference_idle",
|
| 228 |
+
"sub_phase": "req_29"
|
| 229 |
+
},
|
| 230 |
+
{
|
| 231 |
+
"timestamp": "2026-02-16T20:54:57.439263",
|
| 232 |
+
"timestamp_ms": 1771242897439.279,
|
| 233 |
+
"phase": "inference_idle",
|
| 234 |
+
"sub_phase": "req_30"
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"timestamp": "2026-02-16T20:55:02.453135",
|
| 238 |
+
"timestamp_ms": 1771242902453.152,
|
| 239 |
+
"phase": "inference_idle",
|
| 240 |
+
"sub_phase": "req_31"
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"timestamp": "2026-02-16T20:55:07.466953",
|
| 244 |
+
"timestamp_ms": 1771242907466.9663,
|
| 245 |
+
"phase": "inference_idle",
|
| 246 |
+
"sub_phase": "req_32"
|
| 247 |
+
},
|
| 248 |
+
{
|
| 249 |
+
"timestamp": "2026-02-16T20:55:12.480289",
|
| 250 |
+
"timestamp_ms": 1771242912480.3022,
|
| 251 |
+
"phase": "inference_idle",
|
| 252 |
+
"sub_phase": "req_33"
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"timestamp": "2026-02-16T20:55:17.493683",
|
| 256 |
+
"timestamp_ms": 1771242917493.6953,
|
| 257 |
+
"phase": "inference_idle",
|
| 258 |
+
"sub_phase": "req_34"
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"timestamp": "2026-02-16T20:55:22.507032",
|
| 262 |
+
"timestamp_ms": 1771242922507.046,
|
| 263 |
+
"phase": "inference_idle",
|
| 264 |
+
"sub_phase": "req_35"
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"timestamp": "2026-02-16T20:55:27.520514",
|
| 268 |
+
"timestamp_ms": 1771242927520.5305,
|
| 269 |
+
"phase": "inference_idle",
|
| 270 |
+
"sub_phase": "req_36"
|
| 271 |
+
},
|
| 272 |
+
{
|
| 273 |
+
"timestamp": "2026-02-16T20:55:32.534124",
|
| 274 |
+
"timestamp_ms": 1771242932534.1401,
|
| 275 |
+
"phase": "inference_idle",
|
| 276 |
+
"sub_phase": "req_37"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"timestamp": "2026-02-16T20:55:37.547447",
|
| 280 |
+
"timestamp_ms": 1771242937547.4604,
|
| 281 |
+
"phase": "inference_idle",
|
| 282 |
+
"sub_phase": "req_38"
|
| 283 |
+
},
|
| 284 |
+
{
|
| 285 |
+
"timestamp": "2026-02-16T20:55:42.560785",
|
| 286 |
+
"timestamp_ms": 1771242942560.7979,
|
| 287 |
+
"phase": "inference_idle",
|
| 288 |
+
"sub_phase": "req_39"
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"timestamp": "2026-02-16T20:55:47.574025",
|
| 292 |
+
"timestamp_ms": 1771242947574.0376,
|
| 293 |
+
"phase": "inference_idle",
|
| 294 |
+
"sub_phase": "req_40"
|
| 295 |
+
},
|
| 296 |
+
{
|
| 297 |
+
"timestamp": "2026-02-16T20:55:52.587582",
|
| 298 |
+
"timestamp_ms": 1771242952587.5996,
|
| 299 |
+
"phase": "inference_idle",
|
| 300 |
+
"sub_phase": "req_41"
|
| 301 |
+
},
|
| 302 |
+
{
|
| 303 |
+
"timestamp": "2026-02-16T20:55:57.601254",
|
| 304 |
+
"timestamp_ms": 1771242957601.2666,
|
| 305 |
+
"phase": "inference_idle",
|
| 306 |
+
"sub_phase": "req_42"
|
| 307 |
+
},
|
| 308 |
+
{
|
| 309 |
+
"timestamp": "2026-02-16T20:56:02.614434",
|
| 310 |
+
"timestamp_ms": 1771242962614.447,
|
| 311 |
+
"phase": "inference_idle",
|
| 312 |
+
"sub_phase": "req_43"
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"timestamp": "2026-02-16T20:56:07.627974",
|
| 316 |
+
"timestamp_ms": 1771242967627.9873,
|
| 317 |
+
"phase": "inference_idle",
|
| 318 |
+
"sub_phase": "req_44"
|
| 319 |
+
},
|
| 320 |
+
{
|
| 321 |
+
"timestamp": "2026-02-16T20:56:12.641136",
|
| 322 |
+
"timestamp_ms": 1771242972641.1487,
|
| 323 |
+
"phase": "inference_idle",
|
| 324 |
+
"sub_phase": "req_45"
|
| 325 |
+
},
|
| 326 |
+
{
|
| 327 |
+
"timestamp": "2026-02-16T20:56:17.654584",
|
| 328 |
+
"timestamp_ms": 1771242977654.6003,
|
| 329 |
+
"phase": "inference_idle",
|
| 330 |
+
"sub_phase": "req_46"
|
| 331 |
+
},
|
| 332 |
+
{
|
| 333 |
+
"timestamp": "2026-02-16T20:56:22.668100",
|
| 334 |
+
"timestamp_ms": 1771242982668.1162,
|
| 335 |
+
"phase": "inference_idle",
|
| 336 |
+
"sub_phase": "req_47"
|
| 337 |
+
},
|
| 338 |
+
{
|
| 339 |
+
"timestamp": "2026-02-16T20:56:27.681063",
|
| 340 |
+
"timestamp_ms": 1771242987681.0764,
|
| 341 |
+
"phase": "inference_idle",
|
| 342 |
+
"sub_phase": "req_48"
|
| 343 |
+
},
|
| 344 |
+
{
|
| 345 |
+
"timestamp": "2026-02-16T20:56:32.694935",
|
| 346 |
+
"timestamp_ms": 1771242992694.9475,
|
| 347 |
+
"phase": "inference_idle",
|
| 348 |
+
"sub_phase": "req_49"
|
| 349 |
+
},
|
| 350 |
+
{
|
| 351 |
+
"timestamp": "2026-02-16T20:56:37.708628",
|
| 352 |
+
"timestamp_ms": 1771242997708.6443,
|
| 353 |
+
"phase": "inference_idle",
|
| 354 |
+
"sub_phase": "req_50"
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"timestamp": "2026-02-16T20:56:42.721942",
|
| 358 |
+
"timestamp_ms": 1771243002721.9578,
|
| 359 |
+
"phase": "inference_idle",
|
| 360 |
+
"sub_phase": "req_51"
|
| 361 |
+
},
|
| 362 |
+
{
|
| 363 |
+
"timestamp": "2026-02-16T20:56:47.735320",
|
| 364 |
+
"timestamp_ms": 1771243007735.3357,
|
| 365 |
+
"phase": "inference_idle",
|
| 366 |
+
"sub_phase": "req_52"
|
| 367 |
+
},
|
| 368 |
+
{
|
| 369 |
+
"timestamp": "2026-02-16T20:56:52.749099",
|
| 370 |
+
"timestamp_ms": 1771243012749.1152,
|
| 371 |
+
"phase": "inference_idle",
|
| 372 |
+
"sub_phase": "req_53"
|
| 373 |
+
},
|
| 374 |
+
{
|
| 375 |
+
"timestamp": "2026-02-16T20:56:57.762663",
|
| 376 |
+
"timestamp_ms": 1771243017762.678,
|
| 377 |
+
"phase": "inference_idle",
|
| 378 |
+
"sub_phase": "req_54"
|
| 379 |
+
},
|
| 380 |
+
{
|
| 381 |
+
"timestamp": "2026-02-16T20:57:02.776292",
|
| 382 |
+
"timestamp_ms": 1771243022776.3054,
|
| 383 |
+
"phase": "inference_idle",
|
| 384 |
+
"sub_phase": "req_55"
|
| 385 |
+
},
|
| 386 |
+
{
|
| 387 |
+
"timestamp": "2026-02-16T20:57:07.789653",
|
| 388 |
+
"timestamp_ms": 1771243027789.6694,
|
| 389 |
+
"phase": "inference_idle",
|
| 390 |
+
"sub_phase": "req_56"
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"timestamp": "2026-02-16T20:57:12.804018",
|
| 394 |
+
"timestamp_ms": 1771243032804.0396,
|
| 395 |
+
"phase": "inference_idle",
|
| 396 |
+
"sub_phase": "req_57"
|
| 397 |
+
},
|
| 398 |
+
{
|
| 399 |
+
"timestamp": "2026-02-16T20:57:17.817559",
|
| 400 |
+
"timestamp_ms": 1771243037817.575,
|
| 401 |
+
"phase": "inference_idle",
|
| 402 |
+
"sub_phase": "req_58"
|
| 403 |
+
},
|
| 404 |
+
{
|
| 405 |
+
"timestamp": "2026-02-16T20:57:22.830944",
|
| 406 |
+
"timestamp_ms": 1771243042830.96,
|
| 407 |
+
"phase": "inference_idle",
|
| 408 |
+
"sub_phase": "req_59"
|
| 409 |
+
},
|
| 410 |
+
{
|
| 411 |
+
"timestamp": "2026-02-16T20:57:27.831618",
|
| 412 |
+
"timestamp_ms": 1771243047831.6543,
|
| 413 |
+
"phase": "inference",
|
| 414 |
+
"sub_phase": "finished"
|
| 415 |
+
},
|
| 416 |
+
{
|
| 417 |
+
"timestamp": "2026-02-16T20:57:27.831877",
|
| 418 |
+
"timestamp_ms": 1771243047831.8867,
|
| 419 |
+
"phase": "idle_post",
|
| 420 |
+
"sub_phase": ""
|
| 421 |
+
}
|
| 422 |
+
]
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_metrics.png
ADDED
|
Git LFS Details
|
outputs/gpu_power_experiment/06_model_scaling_image/fixed_swin_t_cifar10_gpu0_bs64_ep1_req10000_nocap_fixed/plots/gpu_power_interactive.html
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|