Amshaker commited on
Commit
e46b109
·
verified ·
1 Parent(s): 8c4e79a

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -159,3 +159,4 @@ bimamba_stage_a/wandb/offline-run-20260227_141732-lxvmjzq8/run-lxvmjzq8.wandb fi
159
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
160
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
161
  bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/run-cbyk5tna.wandb filter=lfs diff=lfs merge=lfs -text
 
 
159
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
160
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
161
  bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/run-cbyk5tna.wandb filter=lfs diff=lfs merge=lfs -text
162
+ bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/run-pw3912fk.wandb filter=lfs diff=lfs merge=lfs -text
bimamba_stage_a_40k/checkpoints/epoch2_step28000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2788ff9473de59b89fb0f72a3a55bc5f722de3d928e6fefc0157d7788c1fda0
3
+ size 3418025042
bimamba_stage_a_40k/wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
bimamba_stage_a_40k/wandb/debug.log ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Configure stats pid to 3444203
3
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from /home/x_fahkh/.config/wandb/settings
4
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from /proj/cvl/users/x_fahkh2/WorldMem_Repro/wandb/settings
5
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from environment variables: {'disabled': 'true'}
6
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Applying setup settings: {'_disable_service': False}
7
+ 2026-03-21 14:07:31,110 WARNING MainThread:3444203 [wandb_setup.py:_flush():77] Could not find program at -m main
8
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m main'}
9
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:_log_setup():524] Logging user logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug.log
10
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:_log_setup():525] Logging internal logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug-internal.log
11
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():608] calling init triggers
12
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
13
+ config: {'experiment': {'debug': '${debug}', 'tasks': ['training'], 'num_nodes': 1, 'training': {'precision': '16-mixed', 'compile': False, 'lr': 0.0002, 'batch_size': 128, 'max_epochs': -1, 'max_steps': 40000, 'max_time': None, 'data': {'num_workers': 4, 'shuffle': True}, 'optim': {'accumulate_grad_batches': 1, 'gradient_clip_val': 1.0}, 'checkpointing': {'every_n_train_steps': 2000, 'every_n_epochs': None, 'train_time_interval': None, 'enable_version_counter': False}}, 'validation': {'precision': '16-mixed', 'compile': False, 'batch_size': 4, 'val_every_n_step': 2000, 'val_every_n_epoch': None, 'limit_batch': 8, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'test': {'precision': '16-mixed', 'compile': False, 'batch_size': 1, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'logging': {'metrics': None}, '_name': 'exp_video'}, 'dataset': {'debug': '${debug}', 'metadata': 'data/${dataset.name}/metadata.json', 'data_mean': 0.5, 'data_std': 0.5, 'save_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft', 'n_frames': 300, 'context_length': 1, 'resolution': 128, 'observation_shape': [3, '${dataset.resolution}', '${dataset.resolution}'], 'external_cond_dim': 0, 'validation_multiplier': 1, 'frame_skip': 1, 'action_cond_dim': 25, '_name': 'video_minecraft_latent', 'precomputed_feature_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features', 'use_explicit_memory_frames': False, 'n_frames_valid': 300, 'angle_range': 110, 'pos_range': 8, 'wo_updown': False, 'customized_validation': True, 'add_timestamp_embedding': True}, 'algorithm': {'debug': '${debug}', 'lr': '${experiment.training.lr}', 'x_shape': '${dataset.observation_shape}', 'frame_stack': 1, 'frame_skip': '${dataset.frame_skip}', 'data_mean': '${dataset.data_mean}', 'data_std': '${dataset.data_std}', 'external_cond_dim': 0, 'context_frames': 100, 'weight_decay': 0.002, 'warmup_steps': 1000, 'optimizer_beta': [0.9, 0.99], 'uncertainty_scale': 1, 'guidance_scale': 0.0, 'chunk_size': 1, 'scheduling_matrix': 'autoregressive', 'noise_level': 'random_all', 'causal': True, 'diffusion': {'objective': 'pred_v', 'beta_schedule': 'sigmoid', 'schedule_fn_kwargs': {}, 'clip_noise': 20.0, 'use_snr': False, 'use_cum_snr': False, 'use_fused_snr': True, 'snr_clip': 5.0, 'cum_snr_decay': 0.96, 'timesteps': 1000, 'sampling_timesteps': 20, 'ddim_sampling_eta': 0.0, 'stabilization_level': 15, 'architecture': {'network_size': 64, 'attn_heads': 4, 'attn_dim_head': 64, 'dim_mults': [1, 2, 4, 8], 'resolution': '${dataset.resolution}', 'attn_resolutions': [16, 32, 64, 128], 'use_init_temporal_attn': True, 'use_linear_attn': True, 'time_emb_type': 'rotary'}}, 'n_frames': '${dataset.n_frames}', 'metadata': '${dataset.metadata}', 'action_cond_dim': 25, 'use_plucker': True, 'memory_condition_length': 0, 'log_video': True, 'use_mamba_memory_pipeline': True, 'training_stage': 'stage_a_memory_pretrain', 'stage_b_joint_training': False, 'stage_b_memory_aux_weight': 0.1, 'diff_window_size': 8, 'memory_gap_aux_weight': 0.1, 'memory_gap_aux_weight_init': 0.1, 'memory_gap_aux_weight_final': 0.0, 'memory_gap_aux_decay_fraction': 0.3, 'memory_gap_aux_decay_steps': 0, 'memory_gap_aux_exp_k': 2.0, 'use_precomputed_features': True, 'mamba_latent_channels': 16, 'mamba_model_dim': 256, 'mamba_depth': 4, 'mamba_cond_dim': 256, 'mamba_d_state': 16, 'mamba_d_conv': 4, 'mamba_expand': 2, 'allow_mamba_fallback': False, 'strict_causal_training': True, 'strict_causal_evaluation': True, 'use_oracle_pose_eval': True, 'enable_memory_noise_curriculum': False, 'curriculum_phase_boundaries': [0.2, 0.7], 'curriculum_noise_ranges': [[600, 1000], [200, 900], [0, 400]], 'curriculum_horizons': [50, 100, 200], '_name': 'df_video_mamba3stage', 'use_memory_attention': False, 'relative_embedding': False, 'memory_retrieval_topk': 32}, 'debug': False, 'wandb': {'entity': 'turlin', 'project': 'worldmem', 'mode': 'online'}, 'resume': None, 'load': None, 'name': 'train_stage_a_mamba', 'customized_load': True, 'seperate_load': True, 'diffusion_model_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt', 'vae_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt', 'output_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/'}
14
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():658] starting backend
15
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():662] setting up manager
16
+ 2026-03-21 14:07:31,129 INFO MainThread:3444203 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-03-21 14:07:31,138 INFO MainThread:3444203 [wandb_init.py:init():670] backend started and connected
18
+ 2026-03-21 14:07:31,753 INFO MainThread:3444203 [wandb_init.py:init():768] updated telemetry
19
+ 2026-03-21 14:07:31,817 INFO MainThread:3444203 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
20
+ 2026-03-21 14:07:31,821 INFO MainThread:3444203 [wandb_init.py:init():852] starting run threads in backend
21
+ 2026-03-21 14:07:39,294 INFO MainThread:3444203 [wandb_run.py:_console_start():2465] atexit reg
22
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2026-03-21 14:07:39,309 INFO MainThread:3444203 [wandb_init.py:init():895] run started, returning control to user process
bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/files/conda-environment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ name: worldmem
2
+ channels:
3
+ - defaults
4
+ prefix: /home/x_fahkh/.conda/envs/worldmem
bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/files/wandb-metadata.json ADDED
@@ -0,0 +1,1234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.14.0-611.38.1.el9_7.x86_64-x86_64-with-glibc2.34",
3
+ "python": "3.10.19",
4
+ "heartbeatAt": "2026-03-21T13:07:31.971842",
5
+ "startedAt": "2026-03-21T13:07:31.087391",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "+name=train_stage_a_mamba",
10
+ "algorithm=df_video_mamba3stage",
11
+ "+customized_load=true",
12
+ "+seperate_load=true",
13
+ "+diffusion_model_path=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt",
14
+ "+vae_path=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt",
15
+ "dataset=video_minecraft_latent",
16
+ "dataset.precomputed_feature_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features",
17
+ "dataset.save_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft",
18
+ "dataset.n_frames=300",
19
+ "+dataset.n_frames_valid=300",
20
+ "+dataset.angle_range=110",
21
+ "+dataset.pos_range=8",
22
+ "+dataset.wo_updown=false",
23
+ "+dataset.customized_validation=true",
24
+ "+dataset.add_timestamp_embedding=true",
25
+ "dataset.use_explicit_memory_frames=false",
26
+ "algorithm.training_stage=stage_a_memory_pretrain",
27
+ "algorithm.use_mamba_memory_pipeline=true",
28
+ "algorithm.use_oracle_pose_eval=true",
29
+ "algorithm.use_precomputed_features=true",
30
+ "algorithm.enable_memory_noise_curriculum=false",
31
+ "algorithm.context_frames=100",
32
+ "+algorithm.use_memory_attention=false",
33
+ "+algorithm.relative_embedding=false",
34
+ "+algorithm.memory_retrieval_topk=32",
35
+ "algorithm.memory_gap_aux_weight=0.1",
36
+ "algorithm.memory_gap_aux_weight_init=0.1",
37
+ "algorithm.memory_gap_aux_weight_final=0.0",
38
+ "algorithm.memory_gap_aux_decay_fraction=0.3",
39
+ "algorithm.memory_gap_aux_exp_k=2.0",
40
+ "experiment.training.lr=2e-4",
41
+ "experiment.training.batch_size=128",
42
+ "experiment.training.checkpointing.every_n_train_steps=2000",
43
+ "experiment.training.max_steps=40000",
44
+ "experiment.validation.val_every_n_step=2000",
45
+ "experiment.validation.limit_batch=8",
46
+ "+output_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/"
47
+ ],
48
+ "state": "running",
49
+ "program": "-m main",
50
+ "codePathLocal": null,
51
+ "git": {
52
+ "remote": "https://huggingface.co/BonanDing/WorldMem_Repro",
53
+ "commit": "f14a639b3666f1618923a549415d68fdc13c09d5"
54
+ },
55
+ "email": "fahad.khan@liu.se",
56
+ "root": "/proj/cvl/users/x_fahkh2/WorldMem_Repro",
57
+ "host": "node101",
58
+ "username": "x_fahkh",
59
+ "executable": "/proj/cvl/users/x_fahkh2/envs/worldmem/bin/python",
60
+ "cpu_count": 112,
61
+ "cpu_count_logical": 224,
62
+ "cpu_freq": {
63
+ "current": 1112.3505,
64
+ "min": 800.0,
65
+ "max": 3800.0
66
+ },
67
+ "cpu_freq_per_core": [
68
+ {
69
+ "current": 3770.476,
70
+ "min": 800.0,
71
+ "max": 3800.0
72
+ },
73
+ {
74
+ "current": 800.0,
75
+ "min": 800.0,
76
+ "max": 3800.0
77
+ },
78
+ {
79
+ "current": 3800.0,
80
+ "min": 800.0,
81
+ "max": 3800.0
82
+ },
83
+ {
84
+ "current": 3800.038,
85
+ "min": 800.0,
86
+ "max": 3800.0
87
+ },
88
+ {
89
+ "current": 800.0,
90
+ "min": 800.0,
91
+ "max": 3800.0
92
+ },
93
+ {
94
+ "current": 800.0,
95
+ "min": 800.0,
96
+ "max": 3800.0
97
+ },
98
+ {
99
+ "current": 800.0,
100
+ "min": 800.0,
101
+ "max": 3800.0
102
+ },
103
+ {
104
+ "current": 800.0,
105
+ "min": 800.0,
106
+ "max": 3800.0
107
+ },
108
+ {
109
+ "current": 800.0,
110
+ "min": 800.0,
111
+ "max": 3800.0
112
+ },
113
+ {
114
+ "current": 800.0,
115
+ "min": 800.0,
116
+ "max": 3800.0
117
+ },
118
+ {
119
+ "current": 800.0,
120
+ "min": 800.0,
121
+ "max": 3800.0
122
+ },
123
+ {
124
+ "current": 800.0,
125
+ "min": 800.0,
126
+ "max": 3800.0
127
+ },
128
+ {
129
+ "current": 800.0,
130
+ "min": 800.0,
131
+ "max": 3800.0
132
+ },
133
+ {
134
+ "current": 800.0,
135
+ "min": 800.0,
136
+ "max": 3800.0
137
+ },
138
+ {
139
+ "current": 800.0,
140
+ "min": 800.0,
141
+ "max": 3800.0
142
+ },
143
+ {
144
+ "current": 800.0,
145
+ "min": 800.0,
146
+ "max": 3800.0
147
+ },
148
+ {
149
+ "current": 800.0,
150
+ "min": 800.0,
151
+ "max": 3800.0
152
+ },
153
+ {
154
+ "current": 800.0,
155
+ "min": 800.0,
156
+ "max": 3800.0
157
+ },
158
+ {
159
+ "current": 800.0,
160
+ "min": 800.0,
161
+ "max": 3800.0
162
+ },
163
+ {
164
+ "current": 800.0,
165
+ "min": 800.0,
166
+ "max": 3800.0
167
+ },
168
+ {
169
+ "current": 800.0,
170
+ "min": 800.0,
171
+ "max": 3800.0
172
+ },
173
+ {
174
+ "current": 800.0,
175
+ "min": 800.0,
176
+ "max": 3800.0
177
+ },
178
+ {
179
+ "current": 800.0,
180
+ "min": 800.0,
181
+ "max": 3800.0
182
+ },
183
+ {
184
+ "current": 800.0,
185
+ "min": 800.0,
186
+ "max": 3800.0
187
+ },
188
+ {
189
+ "current": 800.0,
190
+ "min": 800.0,
191
+ "max": 3800.0
192
+ },
193
+ {
194
+ "current": 800.0,
195
+ "min": 800.0,
196
+ "max": 3800.0
197
+ },
198
+ {
199
+ "current": 3799.961,
200
+ "min": 800.0,
201
+ "max": 3800.0
202
+ },
203
+ {
204
+ "current": 800.0,
205
+ "min": 800.0,
206
+ "max": 3800.0
207
+ },
208
+ {
209
+ "current": 3799.916,
210
+ "min": 800.0,
211
+ "max": 3800.0
212
+ },
213
+ {
214
+ "current": 800.0,
215
+ "min": 800.0,
216
+ "max": 3800.0
217
+ },
218
+ {
219
+ "current": 800.0,
220
+ "min": 800.0,
221
+ "max": 3800.0
222
+ },
223
+ {
224
+ "current": 800.0,
225
+ "min": 800.0,
226
+ "max": 3800.0
227
+ },
228
+ {
229
+ "current": 800.0,
230
+ "min": 800.0,
231
+ "max": 3800.0
232
+ },
233
+ {
234
+ "current": 800.0,
235
+ "min": 800.0,
236
+ "max": 3800.0
237
+ },
238
+ {
239
+ "current": 800.0,
240
+ "min": 800.0,
241
+ "max": 3800.0
242
+ },
243
+ {
244
+ "current": 800.0,
245
+ "min": 800.0,
246
+ "max": 3800.0
247
+ },
248
+ {
249
+ "current": 800.0,
250
+ "min": 800.0,
251
+ "max": 3800.0
252
+ },
253
+ {
254
+ "current": 800.0,
255
+ "min": 800.0,
256
+ "max": 3800.0
257
+ },
258
+ {
259
+ "current": 800.0,
260
+ "min": 800.0,
261
+ "max": 3800.0
262
+ },
263
+ {
264
+ "current": 800.0,
265
+ "min": 800.0,
266
+ "max": 3800.0
267
+ },
268
+ {
269
+ "current": 800.0,
270
+ "min": 800.0,
271
+ "max": 3800.0
272
+ },
273
+ {
274
+ "current": 800.0,
275
+ "min": 800.0,
276
+ "max": 3800.0
277
+ },
278
+ {
279
+ "current": 800.0,
280
+ "min": 800.0,
281
+ "max": 3800.0
282
+ },
283
+ {
284
+ "current": 800.0,
285
+ "min": 800.0,
286
+ "max": 3800.0
287
+ },
288
+ {
289
+ "current": 800.0,
290
+ "min": 800.0,
291
+ "max": 3800.0
292
+ },
293
+ {
294
+ "current": 800.0,
295
+ "min": 800.0,
296
+ "max": 3800.0
297
+ },
298
+ {
299
+ "current": 800.0,
300
+ "min": 800.0,
301
+ "max": 3800.0
302
+ },
303
+ {
304
+ "current": 800.0,
305
+ "min": 800.0,
306
+ "max": 3800.0
307
+ },
308
+ {
309
+ "current": 800.0,
310
+ "min": 800.0,
311
+ "max": 3800.0
312
+ },
313
+ {
314
+ "current": 800.0,
315
+ "min": 800.0,
316
+ "max": 3800.0
317
+ },
318
+ {
319
+ "current": 800.0,
320
+ "min": 800.0,
321
+ "max": 3800.0
322
+ },
323
+ {
324
+ "current": 800.0,
325
+ "min": 800.0,
326
+ "max": 3800.0
327
+ },
328
+ {
329
+ "current": 800.0,
330
+ "min": 800.0,
331
+ "max": 3800.0
332
+ },
333
+ {
334
+ "current": 800.0,
335
+ "min": 800.0,
336
+ "max": 3800.0
337
+ },
338
+ {
339
+ "current": 800.0,
340
+ "min": 800.0,
341
+ "max": 3800.0
342
+ },
343
+ {
344
+ "current": 800.0,
345
+ "min": 800.0,
346
+ "max": 3800.0
347
+ },
348
+ {
349
+ "current": 3800.0,
350
+ "min": 800.0,
351
+ "max": 3800.0
352
+ },
353
+ {
354
+ "current": 800.0,
355
+ "min": 800.0,
356
+ "max": 3800.0
357
+ },
358
+ {
359
+ "current": 800.0,
360
+ "min": 800.0,
361
+ "max": 3800.0
362
+ },
363
+ {
364
+ "current": 800.0,
365
+ "min": 800.0,
366
+ "max": 3800.0
367
+ },
368
+ {
369
+ "current": 800.0,
370
+ "min": 800.0,
371
+ "max": 3800.0
372
+ },
373
+ {
374
+ "current": 3798.167,
375
+ "min": 800.0,
376
+ "max": 3800.0
377
+ },
378
+ {
379
+ "current": 800.0,
380
+ "min": 800.0,
381
+ "max": 3800.0
382
+ },
383
+ {
384
+ "current": 800.0,
385
+ "min": 800.0,
386
+ "max": 3800.0
387
+ },
388
+ {
389
+ "current": 800.0,
390
+ "min": 800.0,
391
+ "max": 3800.0
392
+ },
393
+ {
394
+ "current": 800.0,
395
+ "min": 800.0,
396
+ "max": 3800.0
397
+ },
398
+ {
399
+ "current": 800.0,
400
+ "min": 800.0,
401
+ "max": 3800.0
402
+ },
403
+ {
404
+ "current": 800.0,
405
+ "min": 800.0,
406
+ "max": 3800.0
407
+ },
408
+ {
409
+ "current": 800.0,
410
+ "min": 800.0,
411
+ "max": 3800.0
412
+ },
413
+ {
414
+ "current": 800.0,
415
+ "min": 800.0,
416
+ "max": 3800.0
417
+ },
418
+ {
419
+ "current": 800.0,
420
+ "min": 800.0,
421
+ "max": 3800.0
422
+ },
423
+ {
424
+ "current": 800.0,
425
+ "min": 800.0,
426
+ "max": 3800.0
427
+ },
428
+ {
429
+ "current": 800.0,
430
+ "min": 800.0,
431
+ "max": 3800.0
432
+ },
433
+ {
434
+ "current": 800.0,
435
+ "min": 800.0,
436
+ "max": 3800.0
437
+ },
438
+ {
439
+ "current": 800.0,
440
+ "min": 800.0,
441
+ "max": 3800.0
442
+ },
443
+ {
444
+ "current": 800.0,
445
+ "min": 800.0,
446
+ "max": 3800.0
447
+ },
448
+ {
449
+ "current": 3800.038,
450
+ "min": 800.0,
451
+ "max": 3800.0
452
+ },
453
+ {
454
+ "current": 800.0,
455
+ "min": 800.0,
456
+ "max": 3800.0
457
+ },
458
+ {
459
+ "current": 800.0,
460
+ "min": 800.0,
461
+ "max": 3800.0
462
+ },
463
+ {
464
+ "current": 800.0,
465
+ "min": 800.0,
466
+ "max": 3800.0
467
+ },
468
+ {
469
+ "current": 800.0,
470
+ "min": 800.0,
471
+ "max": 3800.0
472
+ },
473
+ {
474
+ "current": 800.0,
475
+ "min": 800.0,
476
+ "max": 3800.0
477
+ },
478
+ {
479
+ "current": 800.0,
480
+ "min": 800.0,
481
+ "max": 3800.0
482
+ },
483
+ {
484
+ "current": 800.0,
485
+ "min": 800.0,
486
+ "max": 3800.0
487
+ },
488
+ {
489
+ "current": 800.0,
490
+ "min": 800.0,
491
+ "max": 3800.0
492
+ },
493
+ {
494
+ "current": 800.0,
495
+ "min": 800.0,
496
+ "max": 3800.0
497
+ },
498
+ {
499
+ "current": 800.0,
500
+ "min": 800.0,
501
+ "max": 3800.0
502
+ },
503
+ {
504
+ "current": 3800.0,
505
+ "min": 800.0,
506
+ "max": 3800.0
507
+ },
508
+ {
509
+ "current": 3795.459,
510
+ "min": 800.0,
511
+ "max": 3800.0
512
+ },
513
+ {
514
+ "current": 800.0,
515
+ "min": 800.0,
516
+ "max": 3800.0
517
+ },
518
+ {
519
+ "current": 800.0,
520
+ "min": 800.0,
521
+ "max": 3800.0
522
+ },
523
+ {
524
+ "current": 3785.016,
525
+ "min": 800.0,
526
+ "max": 3800.0
527
+ },
528
+ {
529
+ "current": 800.0,
530
+ "min": 800.0,
531
+ "max": 3800.0
532
+ },
533
+ {
534
+ "current": 3804.352,
535
+ "min": 800.0,
536
+ "max": 3800.0
537
+ },
538
+ {
539
+ "current": 800.0,
540
+ "min": 800.0,
541
+ "max": 3800.0
542
+ },
543
+ {
544
+ "current": 800.0,
545
+ "min": 800.0,
546
+ "max": 3800.0
547
+ },
548
+ {
549
+ "current": 800.0,
550
+ "min": 800.0,
551
+ "max": 3800.0
552
+ },
553
+ {
554
+ "current": 800.0,
555
+ "min": 800.0,
556
+ "max": 3800.0
557
+ },
558
+ {
559
+ "current": 800.0,
560
+ "min": 800.0,
561
+ "max": 3800.0
562
+ },
563
+ {
564
+ "current": 3692.474,
565
+ "min": 800.0,
566
+ "max": 3800.0
567
+ },
568
+ {
569
+ "current": 800.0,
570
+ "min": 800.0,
571
+ "max": 3800.0
572
+ },
573
+ {
574
+ "current": 800.0,
575
+ "min": 800.0,
576
+ "max": 3800.0
577
+ },
578
+ {
579
+ "current": 800.0,
580
+ "min": 800.0,
581
+ "max": 3800.0
582
+ },
583
+ {
584
+ "current": 800.0,
585
+ "min": 800.0,
586
+ "max": 3800.0
587
+ },
588
+ {
589
+ "current": 800.0,
590
+ "min": 800.0,
591
+ "max": 3800.0
592
+ },
593
+ {
594
+ "current": 800.0,
595
+ "min": 800.0,
596
+ "max": 3800.0
597
+ },
598
+ {
599
+ "current": 800.0,
600
+ "min": 800.0,
601
+ "max": 3800.0
602
+ },
603
+ {
604
+ "current": 800.0,
605
+ "min": 800.0,
606
+ "max": 3800.0
607
+ },
608
+ {
609
+ "current": 800.0,
610
+ "min": 800.0,
611
+ "max": 3800.0
612
+ },
613
+ {
614
+ "current": 800.0,
615
+ "min": 800.0,
616
+ "max": 3800.0
617
+ },
618
+ {
619
+ "current": 800.0,
620
+ "min": 800.0,
621
+ "max": 3800.0
622
+ },
623
+ {
624
+ "current": 800.0,
625
+ "min": 800.0,
626
+ "max": 3800.0
627
+ },
628
+ {
629
+ "current": 800.0,
630
+ "min": 800.0,
631
+ "max": 3800.0
632
+ },
633
+ {
634
+ "current": 800.0,
635
+ "min": 800.0,
636
+ "max": 3800.0
637
+ },
638
+ {
639
+ "current": 800.0,
640
+ "min": 800.0,
641
+ "max": 3800.0
642
+ },
643
+ {
644
+ "current": 800.0,
645
+ "min": 800.0,
646
+ "max": 3800.0
647
+ },
648
+ {
649
+ "current": 800.0,
650
+ "min": 800.0,
651
+ "max": 3800.0
652
+ },
653
+ {
654
+ "current": 3802.128,
655
+ "min": 800.0,
656
+ "max": 3800.0
657
+ },
658
+ {
659
+ "current": 800.0,
660
+ "min": 800.0,
661
+ "max": 3800.0
662
+ },
663
+ {
664
+ "current": 3572.978,
665
+ "min": 800.0,
666
+ "max": 3800.0
667
+ },
668
+ {
669
+ "current": 800.0,
670
+ "min": 800.0,
671
+ "max": 3800.0
672
+ },
673
+ {
674
+ "current": 800.0,
675
+ "min": 800.0,
676
+ "max": 3800.0
677
+ },
678
+ {
679
+ "current": 800.0,
680
+ "min": 800.0,
681
+ "max": 3800.0
682
+ },
683
+ {
684
+ "current": 800.0,
685
+ "min": 800.0,
686
+ "max": 3800.0
687
+ },
688
+ {
689
+ "current": 800.0,
690
+ "min": 800.0,
691
+ "max": 3800.0
692
+ },
693
+ {
694
+ "current": 800.0,
695
+ "min": 800.0,
696
+ "max": 3800.0
697
+ },
698
+ {
699
+ "current": 800.0,
700
+ "min": 800.0,
701
+ "max": 3800.0
702
+ },
703
+ {
704
+ "current": 800.0,
705
+ "min": 800.0,
706
+ "max": 3800.0
707
+ },
708
+ {
709
+ "current": 800.0,
710
+ "min": 800.0,
711
+ "max": 3800.0
712
+ },
713
+ {
714
+ "current": 800.0,
715
+ "min": 800.0,
716
+ "max": 3800.0
717
+ },
718
+ {
719
+ "current": 3800.0,
720
+ "min": 800.0,
721
+ "max": 3800.0
722
+ },
723
+ {
724
+ "current": 800.0,
725
+ "min": 800.0,
726
+ "max": 3800.0
727
+ },
728
+ {
729
+ "current": 800.0,
730
+ "min": 800.0,
731
+ "max": 3800.0
732
+ },
733
+ {
734
+ "current": 800.0,
735
+ "min": 800.0,
736
+ "max": 3800.0
737
+ },
738
+ {
739
+ "current": 800.0,
740
+ "min": 800.0,
741
+ "max": 3800.0
742
+ },
743
+ {
744
+ "current": 800.0,
745
+ "min": 800.0,
746
+ "max": 3800.0
747
+ },
748
+ {
749
+ "current": 800.0,
750
+ "min": 800.0,
751
+ "max": 3800.0
752
+ },
753
+ {
754
+ "current": 800.0,
755
+ "min": 800.0,
756
+ "max": 3800.0
757
+ },
758
+ {
759
+ "current": 800.0,
760
+ "min": 800.0,
761
+ "max": 3800.0
762
+ },
763
+ {
764
+ "current": 800.0,
765
+ "min": 800.0,
766
+ "max": 3800.0
767
+ },
768
+ {
769
+ "current": 800.0,
770
+ "min": 800.0,
771
+ "max": 3800.0
772
+ },
773
+ {
774
+ "current": 800.0,
775
+ "min": 800.0,
776
+ "max": 3800.0
777
+ },
778
+ {
779
+ "current": 800.0,
780
+ "min": 800.0,
781
+ "max": 3800.0
782
+ },
783
+ {
784
+ "current": 800.0,
785
+ "min": 800.0,
786
+ "max": 3800.0
787
+ },
788
+ {
789
+ "current": 800.0,
790
+ "min": 800.0,
791
+ "max": 3800.0
792
+ },
793
+ {
794
+ "current": 800.0,
795
+ "min": 800.0,
796
+ "max": 3800.0
797
+ },
798
+ {
799
+ "current": 800.0,
800
+ "min": 800.0,
801
+ "max": 3800.0
802
+ },
803
+ {
804
+ "current": 800.0,
805
+ "min": 800.0,
806
+ "max": 3800.0
807
+ },
808
+ {
809
+ "current": 800.0,
810
+ "min": 800.0,
811
+ "max": 3800.0
812
+ },
813
+ {
814
+ "current": 800.0,
815
+ "min": 800.0,
816
+ "max": 3800.0
817
+ },
818
+ {
819
+ "current": 3800.0,
820
+ "min": 800.0,
821
+ "max": 3800.0
822
+ },
823
+ {
824
+ "current": 800.0,
825
+ "min": 800.0,
826
+ "max": 3800.0
827
+ },
828
+ {
829
+ "current": 800.0,
830
+ "min": 800.0,
831
+ "max": 3800.0
832
+ },
833
+ {
834
+ "current": 800.0,
835
+ "min": 800.0,
836
+ "max": 3800.0
837
+ },
838
+ {
839
+ "current": 800.0,
840
+ "min": 800.0,
841
+ "max": 3800.0
842
+ },
843
+ {
844
+ "current": 800.0,
845
+ "min": 800.0,
846
+ "max": 3800.0
847
+ },
848
+ {
849
+ "current": 800.0,
850
+ "min": 800.0,
851
+ "max": 3800.0
852
+ },
853
+ {
854
+ "current": 800.0,
855
+ "min": 800.0,
856
+ "max": 3800.0
857
+ },
858
+ {
859
+ "current": 800.0,
860
+ "min": 800.0,
861
+ "max": 3800.0
862
+ },
863
+ {
864
+ "current": 800.0,
865
+ "min": 800.0,
866
+ "max": 3800.0
867
+ },
868
+ {
869
+ "current": 800.0,
870
+ "min": 800.0,
871
+ "max": 3800.0
872
+ },
873
+ {
874
+ "current": 800.0,
875
+ "min": 800.0,
876
+ "max": 3800.0
877
+ },
878
+ {
879
+ "current": 800.0,
880
+ "min": 800.0,
881
+ "max": 3800.0
882
+ },
883
+ {
884
+ "current": 800.0,
885
+ "min": 800.0,
886
+ "max": 3800.0
887
+ },
888
+ {
889
+ "current": 3705.778,
890
+ "min": 800.0,
891
+ "max": 3800.0
892
+ },
893
+ {
894
+ "current": 3800.0,
895
+ "min": 800.0,
896
+ "max": 3800.0
897
+ },
898
+ {
899
+ "current": 800.0,
900
+ "min": 800.0,
901
+ "max": 3800.0
902
+ },
903
+ {
904
+ "current": 800.0,
905
+ "min": 800.0,
906
+ "max": 3800.0
907
+ },
908
+ {
909
+ "current": 800.0,
910
+ "min": 800.0,
911
+ "max": 3800.0
912
+ },
913
+ {
914
+ "current": 800.0,
915
+ "min": 800.0,
916
+ "max": 3800.0
917
+ },
918
+ {
919
+ "current": 800.0,
920
+ "min": 800.0,
921
+ "max": 3800.0
922
+ },
923
+ {
924
+ "current": 800.0,
925
+ "min": 800.0,
926
+ "max": 3800.0
927
+ },
928
+ {
929
+ "current": 800.0,
930
+ "min": 800.0,
931
+ "max": 3800.0
932
+ },
933
+ {
934
+ "current": 800.0,
935
+ "min": 800.0,
936
+ "max": 3800.0
937
+ },
938
+ {
939
+ "current": 800.0,
940
+ "min": 800.0,
941
+ "max": 3800.0
942
+ },
943
+ {
944
+ "current": 800.0,
945
+ "min": 800.0,
946
+ "max": 3800.0
947
+ },
948
+ {
949
+ "current": 3799.961,
950
+ "min": 800.0,
951
+ "max": 3800.0
952
+ },
953
+ {
954
+ "current": 800.0,
955
+ "min": 800.0,
956
+ "max": 3800.0
957
+ },
958
+ {
959
+ "current": 800.0,
960
+ "min": 800.0,
961
+ "max": 3800.0
962
+ },
963
+ {
964
+ "current": 800.0,
965
+ "min": 800.0,
966
+ "max": 3800.0
967
+ },
968
+ {
969
+ "current": 800.0,
970
+ "min": 800.0,
971
+ "max": 3800.0
972
+ },
973
+ {
974
+ "current": 800.0,
975
+ "min": 800.0,
976
+ "max": 3800.0
977
+ },
978
+ {
979
+ "current": 800.0,
980
+ "min": 800.0,
981
+ "max": 3800.0
982
+ },
983
+ {
984
+ "current": 800.0,
985
+ "min": 800.0,
986
+ "max": 3800.0
987
+ },
988
+ {
989
+ "current": 800.0,
990
+ "min": 800.0,
991
+ "max": 3800.0
992
+ },
993
+ {
994
+ "current": 800.0,
995
+ "min": 800.0,
996
+ "max": 3800.0
997
+ },
998
+ {
999
+ "current": 800.0,
1000
+ "min": 800.0,
1001
+ "max": 3800.0
1002
+ },
1003
+ {
1004
+ "current": 800.0,
1005
+ "min": 800.0,
1006
+ "max": 3800.0
1007
+ },
1008
+ {
1009
+ "current": 3800.74,
1010
+ "min": 800.0,
1011
+ "max": 3800.0
1012
+ },
1013
+ {
1014
+ "current": 800.0,
1015
+ "min": 800.0,
1016
+ "max": 3800.0
1017
+ },
1018
+ {
1019
+ "current": 3800.038,
1020
+ "min": 800.0,
1021
+ "max": 3800.0
1022
+ },
1023
+ {
1024
+ "current": 800.0,
1025
+ "min": 800.0,
1026
+ "max": 3800.0
1027
+ },
1028
+ {
1029
+ "current": 800.0,
1030
+ "min": 800.0,
1031
+ "max": 3800.0
1032
+ },
1033
+ {
1034
+ "current": 800.0,
1035
+ "min": 800.0,
1036
+ "max": 3800.0
1037
+ },
1038
+ {
1039
+ "current": 800.0,
1040
+ "min": 800.0,
1041
+ "max": 3800.0
1042
+ },
1043
+ {
1044
+ "current": 800.0,
1045
+ "min": 800.0,
1046
+ "max": 3800.0
1047
+ },
1048
+ {
1049
+ "current": 800.0,
1050
+ "min": 800.0,
1051
+ "max": 3800.0
1052
+ },
1053
+ {
1054
+ "current": 800.0,
1055
+ "min": 800.0,
1056
+ "max": 3800.0
1057
+ },
1058
+ {
1059
+ "current": 800.0,
1060
+ "min": 800.0,
1061
+ "max": 3800.0
1062
+ },
1063
+ {
1064
+ "current": 800.0,
1065
+ "min": 800.0,
1066
+ "max": 3800.0
1067
+ },
1068
+ {
1069
+ "current": 800.0,
1070
+ "min": 800.0,
1071
+ "max": 3800.0
1072
+ },
1073
+ {
1074
+ "current": 3804.185,
1075
+ "min": 800.0,
1076
+ "max": 3800.0
1077
+ },
1078
+ {
1079
+ "current": 800.0,
1080
+ "min": 800.0,
1081
+ "max": 3800.0
1082
+ },
1083
+ {
1084
+ "current": 800.0,
1085
+ "min": 800.0,
1086
+ "max": 3800.0
1087
+ },
1088
+ {
1089
+ "current": 3237.529,
1090
+ "min": 800.0,
1091
+ "max": 3800.0
1092
+ },
1093
+ {
1094
+ "current": 800.0,
1095
+ "min": 800.0,
1096
+ "max": 3800.0
1097
+ },
1098
+ {
1099
+ "current": 3800.0,
1100
+ "min": 800.0,
1101
+ "max": 3800.0
1102
+ },
1103
+ {
1104
+ "current": 800.0,
1105
+ "min": 800.0,
1106
+ "max": 3800.0
1107
+ },
1108
+ {
1109
+ "current": 800.0,
1110
+ "min": 800.0,
1111
+ "max": 3800.0
1112
+ },
1113
+ {
1114
+ "current": 800.0,
1115
+ "min": 800.0,
1116
+ "max": 3800.0
1117
+ },
1118
+ {
1119
+ "current": 800.0,
1120
+ "min": 800.0,
1121
+ "max": 3800.0
1122
+ },
1123
+ {
1124
+ "current": 800.0,
1125
+ "min": 800.0,
1126
+ "max": 3800.0
1127
+ },
1128
+ {
1129
+ "current": 800.0,
1130
+ "min": 800.0,
1131
+ "max": 3800.0
1132
+ },
1133
+ {
1134
+ "current": 800.0,
1135
+ "min": 800.0,
1136
+ "max": 3800.0
1137
+ },
1138
+ {
1139
+ "current": 800.0,
1140
+ "min": 800.0,
1141
+ "max": 3800.0
1142
+ },
1143
+ {
1144
+ "current": 800.0,
1145
+ "min": 800.0,
1146
+ "max": 3800.0
1147
+ },
1148
+ {
1149
+ "current": 800.0,
1150
+ "min": 800.0,
1151
+ "max": 3800.0
1152
+ },
1153
+ {
1154
+ "current": 800.0,
1155
+ "min": 800.0,
1156
+ "max": 3800.0
1157
+ },
1158
+ {
1159
+ "current": 800.0,
1160
+ "min": 800.0,
1161
+ "max": 3800.0
1162
+ },
1163
+ {
1164
+ "current": 800.0,
1165
+ "min": 800.0,
1166
+ "max": 3800.0
1167
+ },
1168
+ {
1169
+ "current": 800.0,
1170
+ "min": 800.0,
1171
+ "max": 3800.0
1172
+ },
1173
+ {
1174
+ "current": 3799.961,
1175
+ "min": 800.0,
1176
+ "max": 3800.0
1177
+ },
1178
+ {
1179
+ "current": 800.0,
1180
+ "min": 800.0,
1181
+ "max": 3800.0
1182
+ },
1183
+ {
1184
+ "current": 800.0,
1185
+ "min": 800.0,
1186
+ "max": 3800.0
1187
+ }
1188
+ ],
1189
+ "disk": {
1190
+ "/": {
1191
+ "total": 30.58770751953125,
1192
+ "used": 16.39543914794922
1193
+ }
1194
+ },
1195
+ "gpu": "NVIDIA H200",
1196
+ "gpu_count": 8,
1197
+ "gpu_devices": [
1198
+ {
1199
+ "name": "NVIDIA H200",
1200
+ "memory_total": 150754820096
1201
+ },
1202
+ {
1203
+ "name": "NVIDIA H200",
1204
+ "memory_total": 150754820096
1205
+ },
1206
+ {
1207
+ "name": "NVIDIA H200",
1208
+ "memory_total": 150754820096
1209
+ },
1210
+ {
1211
+ "name": "NVIDIA H200",
1212
+ "memory_total": 150754820096
1213
+ },
1214
+ {
1215
+ "name": "NVIDIA H200",
1216
+ "memory_total": 150754820096
1217
+ },
1218
+ {
1219
+ "name": "NVIDIA H200",
1220
+ "memory_total": 150754820096
1221
+ },
1222
+ {
1223
+ "name": "NVIDIA H200",
1224
+ "memory_total": 150754820096
1225
+ },
1226
+ {
1227
+ "name": "NVIDIA H200",
1228
+ "memory_total": 150754820096
1229
+ }
1230
+ ],
1231
+ "memory": {
1232
+ "total": 2015.5592231750488
1233
+ }
1234
+ }
bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug.log ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Configure stats pid to 3444203
3
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from /home/x_fahkh/.config/wandb/settings
4
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from /proj/cvl/users/x_fahkh2/WorldMem_Repro/wandb/settings
5
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Loading settings from environment variables: {'disabled': 'true'}
6
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Applying setup settings: {'_disable_service': False}
7
+ 2026-03-21 14:07:31,110 WARNING MainThread:3444203 [wandb_setup.py:_flush():77] Could not find program at -m main
8
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m main'}
9
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:_log_setup():524] Logging user logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug.log
10
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:_log_setup():525] Logging internal logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/logs/debug-internal.log
11
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():608] calling init triggers
12
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
13
+ config: {'experiment': {'debug': '${debug}', 'tasks': ['training'], 'num_nodes': 1, 'training': {'precision': '16-mixed', 'compile': False, 'lr': 0.0002, 'batch_size': 128, 'max_epochs': -1, 'max_steps': 40000, 'max_time': None, 'data': {'num_workers': 4, 'shuffle': True}, 'optim': {'accumulate_grad_batches': 1, 'gradient_clip_val': 1.0}, 'checkpointing': {'every_n_train_steps': 2000, 'every_n_epochs': None, 'train_time_interval': None, 'enable_version_counter': False}}, 'validation': {'precision': '16-mixed', 'compile': False, 'batch_size': 4, 'val_every_n_step': 2000, 'val_every_n_epoch': None, 'limit_batch': 8, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'test': {'precision': '16-mixed', 'compile': False, 'batch_size': 1, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'logging': {'metrics': None}, '_name': 'exp_video'}, 'dataset': {'debug': '${debug}', 'metadata': 'data/${dataset.name}/metadata.json', 'data_mean': 0.5, 'data_std': 0.5, 'save_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft', 'n_frames': 300, 'context_length': 1, 'resolution': 128, 'observation_shape': [3, '${dataset.resolution}', '${dataset.resolution}'], 'external_cond_dim': 0, 'validation_multiplier': 1, 'frame_skip': 1, 'action_cond_dim': 25, '_name': 'video_minecraft_latent', 'precomputed_feature_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features', 'use_explicit_memory_frames': False, 'n_frames_valid': 300, 'angle_range': 110, 'pos_range': 8, 'wo_updown': False, 'customized_validation': True, 'add_timestamp_embedding': True}, 'algorithm': {'debug': '${debug}', 'lr': '${experiment.training.lr}', 'x_shape': '${dataset.observation_shape}', 'frame_stack': 1, 'frame_skip': '${dataset.frame_skip}', 'data_mean': '${dataset.data_mean}', 'data_std': '${dataset.data_std}', 'external_cond_dim': 0, 'context_frames': 100, 'weight_decay': 0.002, 'warmup_steps': 1000, 'optimizer_beta': [0.9, 0.99], 'uncertainty_scale': 1, 'guidance_scale': 0.0, 'chunk_size': 1, 'scheduling_matrix': 'autoregressive', 'noise_level': 'random_all', 'causal': True, 'diffusion': {'objective': 'pred_v', 'beta_schedule': 'sigmoid', 'schedule_fn_kwargs': {}, 'clip_noise': 20.0, 'use_snr': False, 'use_cum_snr': False, 'use_fused_snr': True, 'snr_clip': 5.0, 'cum_snr_decay': 0.96, 'timesteps': 1000, 'sampling_timesteps': 20, 'ddim_sampling_eta': 0.0, 'stabilization_level': 15, 'architecture': {'network_size': 64, 'attn_heads': 4, 'attn_dim_head': 64, 'dim_mults': [1, 2, 4, 8], 'resolution': '${dataset.resolution}', 'attn_resolutions': [16, 32, 64, 128], 'use_init_temporal_attn': True, 'use_linear_attn': True, 'time_emb_type': 'rotary'}}, 'n_frames': '${dataset.n_frames}', 'metadata': '${dataset.metadata}', 'action_cond_dim': 25, 'use_plucker': True, 'memory_condition_length': 0, 'log_video': True, 'use_mamba_memory_pipeline': True, 'training_stage': 'stage_a_memory_pretrain', 'stage_b_joint_training': False, 'stage_b_memory_aux_weight': 0.1, 'diff_window_size': 8, 'memory_gap_aux_weight': 0.1, 'memory_gap_aux_weight_init': 0.1, 'memory_gap_aux_weight_final': 0.0, 'memory_gap_aux_decay_fraction': 0.3, 'memory_gap_aux_decay_steps': 0, 'memory_gap_aux_exp_k': 2.0, 'use_precomputed_features': True, 'mamba_latent_channels': 16, 'mamba_model_dim': 256, 'mamba_depth': 4, 'mamba_cond_dim': 256, 'mamba_d_state': 16, 'mamba_d_conv': 4, 'mamba_expand': 2, 'allow_mamba_fallback': False, 'strict_causal_training': True, 'strict_causal_evaluation': True, 'use_oracle_pose_eval': True, 'enable_memory_noise_curriculum': False, 'curriculum_phase_boundaries': [0.2, 0.7], 'curriculum_noise_ranges': [[600, 1000], [200, 900], [0, 400]], 'curriculum_horizons': [50, 100, 200], '_name': 'df_video_mamba3stage', 'use_memory_attention': False, 'relative_embedding': False, 'memory_retrieval_topk': 32}, 'debug': False, 'wandb': {'entity': 'turlin', 'project': 'worldmem', 'mode': 'online'}, 'resume': None, 'load': None, 'name': 'train_stage_a_mamba', 'customized_load': True, 'seperate_load': True, 'diffusion_model_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt', 'vae_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt', 'output_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_40k/'}
14
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():658] starting backend
15
+ 2026-03-21 14:07:31,110 INFO MainThread:3444203 [wandb_init.py:init():662] setting up manager
16
+ 2026-03-21 14:07:31,129 INFO MainThread:3444203 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-03-21 14:07:31,138 INFO MainThread:3444203 [wandb_init.py:init():670] backend started and connected
18
+ 2026-03-21 14:07:31,753 INFO MainThread:3444203 [wandb_init.py:init():768] updated telemetry
19
+ 2026-03-21 14:07:31,817 INFO MainThread:3444203 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
20
+ 2026-03-21 14:07:31,821 INFO MainThread:3444203 [wandb_init.py:init():852] starting run threads in backend
21
+ 2026-03-21 14:07:39,294 INFO MainThread:3444203 [wandb_run.py:_console_start():2465] atexit reg
22
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2026-03-21 14:07:39,295 INFO MainThread:3444203 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2026-03-21 14:07:39,309 INFO MainThread:3444203 [wandb_init.py:init():895] run started, returning control to user process
bimamba_stage_a_40k/wandb/offline-run-20260321_140731-pw3912fk/run-pw3912fk.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a710b6b7a871b492719d2e8697de5bc97a5ee4238a9294fb393d9ed589c9bbd1
3
+ size 37036047
bimamba_stage_a_40k/wandb/wandb-resume.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"run_id": "pw3912fk"}